scispace - formally typeset
Search or ask a question
Journal ArticleDOI

A 0.5 V 55 $\mu \text{W}$ 64 $\times $ 2 Channel Binaural Silicon Cochlea for Event-Driven Stereo-Audio Sensing

22 Sep 2016-IEEE Journal of Solid-state Circuits (IEEE)-Vol. 51, Iss: 11, pp 2554-2569
TL;DR: A 0.5V 55μW 64×2-channel binaural silicon cochlea aiming for ultra-low-power IoE applications like event-driven VAD, sound source localization, speaker identification and primitive speech recognition is presented.
Abstract: This paper presents a $64 \times 2$ channel stereo-audio sensing front end with parallel asynchronous event output inspired by the biological cochlea. Each binaural channel performs feature extraction by analog bandpass filtering, and the filtered signal is encoded into events via asynchronous delta modulation (ADM). The channel central frequencies $f_{0}$ are geometrically scaled across the human hearing range. Two design techniques are highlighted to achieve the high system power efficiency: source-follower-based bandpass filters (BPFs) and asynchronous delta modulation (ADM) with adaptive self-oscillating comparison. The chip was fabricated in 0.18 $\mu \text{m}$ 1P6M CMOS, and occupies an area of $10.5 \times 4.8$ mm2. The core cochlea system operating under a 0.5 V power supply consumes 55 $\mu \text{W}$ at an output rate of 100k event/s. The measured range of $f_{0}$ is from 8 Hz to 20 kHz, and the BPF quality factor ${Q}$ can be tuned from 1 to almost 40. The 1 $\sigma $ mismatch of $f_{0}$ and ${Q}$ between two ears is 3.3% and 15%, respectively, across all channels at ${Q}\approx $ 10. Reconstruction of speech input from the event output of the chip is performed to validate the information integrity in event-domain representation, and vowel discrimination is demonstrated as a simple application using histograms of the output events. This type of silicon cochlea front end targets integration with embedded event-driven processors for low-power smart audio sensing with classification capabilities, such as voice activity detection and speaker identification.
Citations
More filters
Journal ArticleDOI
TL;DR: A brief history of neuromorphic engineering is presented, some of the principal current large-scale projects, their main features, how their approaches are complementary and distinct, their advantages and drawbacks, and highlight the sorts of capabilities that each can deliver to neural modellers are highlighted.
Abstract: Neuromorphic computing covers a diverse range of approaches to information processing all of which demonstrate some degree of neurobiological inspiration that differentiates them from mainstream conventional computing systems. The philosophy behind neuromorphic computing has its origins in the seminal work carried out by Carver Mead at Caltech in the late 1980s. This early work influenced others to carry developments forward, and advances in VLSI technology supported steady growth in the scale and capability of neuromorphic devices. Recently, a number of large-scale neuromorphic projects have emerged, taking the approach to unprecedented scales and capabilities. These large-scale projects are associated with major new funding initiatives for brain-related research, creating a sense that the time and circumstances are right for progress in our understanding of information processing in the brain. In this review we present a brief history of neuromorphic engineering then focus on some of the principal current large-scale projects, their main features, how their approaches are complementary and distinct, their advantages and drawbacks, and highlight the sorts of capabilities that each can deliver to neural modellers.

275 citations


Cites background from "A 0.5 V 55 $\mu \text{W}$ 64 $\time..."

  • ...Recent work at INI includes the development of neuromorphic vision sensors [8], silicon cochlea [9], and mediumscale neuromorphic processors such as the Reconfigurable On-Line Learning Spiking (ROLLS) and cxQuad chips....

    [...]

Journal ArticleDOI
TL;DR: In this article , the authors present a snapshot of the present state of neuromorphic technology and provide an opinion on the challenges and opportunities that the future holds in the major areas of the neuromorphic computing community.
Abstract: Abstract Modern computation based on von Neumann architecture is now a mature cutting-edge science. In the von Neumann architecture, processing and memory units are implemented as separate blocks interchanging data intensively and continuously. This data transfer is responsible for a large part of the power consumption. The next generation computer technology is expected to solve problems at the exascale with 10 18 calculations each second. Even though these future computers will be incredibly powerful, if they are based on von Neumann type architectures, they will consume between 20 and 30 megawatts of power and will not have intrinsic physically built-in capabilities to learn or deal with complex data as our brain does. These needs can be addressed by neuromorphic computing systems which are inspired by the biological concepts of the human brain. This new generation of computers has the potential to be used for the storage and processing of large amounts of digital information with much lower power consumption than conventional processors. Among their potential future applications, an important niche is moving the control from data centers to edge devices. The aim of this roadmap is to present a snapshot of the present state of neuromorphic technology and provide an opinion on the challenges and opportunities that the future holds in the major areas of neuromorphic technology, namely materials, devices, neuromorphic circuits, neuromorphic algorithms, applications, and ethics. The roadmap is a collection of perspectives where leading researchers in the neuromorphic community provide their own view about the current state and the future challenges for each research area. We hope that this roadmap will be a useful resource by providing a concise yet comprehensive introduction to readers outside this field, for those who are just entering the field, as well as providing future perspectives for those who are well established in the neuromorphic computing community.

99 citations

Journal ArticleDOI
TL;DR: A survey of recent works in developing neuromorphic or neuro-inspired hardware systems, focusing on those systems which can either learn from data in an unsupervised or online supervised manner, and present algorithms and architectures developed specially to support on-chip learning.
Abstract: In this paper, we present a survey of recent works in developing neuromorphic or neuro-inspired hardware systems. In particular, we focus on those systems which can either learn from data in an unsupervised or online supervised manner. We present algorithms and architectures developed specially to support on-chip learning. Emphasis is placed on hardware friendly modifications of standard algorithms, such as backpropagation, as well as novel algorithms, such as structural plasticity, developed specially for low-resolution synapses. We cover works related to both spike-based and more traditional non-spike-based algorithms. This is followed by developments in novel devices, such as floating-gate MOS, memristors, and spintronic devices. CMOS circuit innovations for on-chip learning and CMOS interface circuits for post-CMOS devices, such as memristors, are presented. Common architectures, such as crossbar or island style arrays, are discussed, along with their relative merits and demerits. Finally, we present some possible applications of neuromorphic hardware, such as brain–machine interfaces, robotics, etc., and identify future research trends in the field.

90 citations


Cites background from "A 0.5 V 55 $\mu \text{W}$ 64 $\time..."

  • ...To mimic biology more closely, neuromorphic visual [178], [179], auditory [180] and tactile [181] sensors have been developed which operate in an asynchronous, spike based manner similar to human retina, cochlea and mechanoreceptors, respectively....

    [...]

Journal ArticleDOI
01 Dec 2020-PhotoniX
TL;DR: The need and the possibility to conceive a photonic memristor are discussed, a positive outlook on the challenges and opportunities for the ambitious goal of realising the next generation of full-optical neuromorphic hardware is offered.
Abstract: Neuromorphic computing applies concepts extracted from neuroscience to develop devices shaped like neural systems and achieve brain-like capacity and efficiency. In this way, neuromorphic machines, able to learn from the surrounding environment to deduce abstract concepts and to make decisions, promise to start a technological revolution transforming our society and our life. Current electronic implementations of neuromorphic architectures are still far from competing with their biological counterparts in terms of real-time information-processing capabilities, packing density and energy efficiency. A solution to this impasse is represented by the application of photonic principles to the neuromorphic domain creating in this way the field of neuromorphic photonics. This new field combines the advantages of photonics and neuromorphic architectures to build systems with high efficiency, high interconnectivity and high information density, and paves the way to ultrafast, power efficient and low cost and complex signal processing. In this Perspective, we review the rapid development of the neuromorphic computing field both in the electronic and in the photonic domain focusing on the role and the applications of memristors. We discuss the need and the possibility to conceive a photonic memristor and we offer a positive outlook on the challenges and opportunities for the ambitious goal of realising the next generation of full-optical neuromorphic hardware.

87 citations

Journal ArticleDOI
TL;DR: This work investigates the effectiveness of synchronous and asynchronous frame-based features generated using spike count and constant event binning in combination with the use of a recurrent neural network for solving a classification task using N-TIDIGITS18 dataset and proposes a new pre-processing method which applies an exponential kernel on the output cochlea spikes so that the interspike timing information is better preserved.
Abstract: Event-driven neuromorphic spiking sensors such as the silicon retina and the silicon cochlea encode the external sensory stimuli as asynchronous streams of spikes across different channels or pixels. Combining state-of-art deep neural networks with the asynchronous outputs of these sensors has produced encouraging results on some datasets but remains challenging. While the lack of effective spiking networks to process the spike streams is one reason, the other reason is that the pre-processing methods required to convert the spike streams to frame-based features needed for the deep networks still require further investigation. This work investigates the effectiveness of synchronous and asynchronous frame-based features generated using spike count and constant event binning in combination with the use of a recurrent neural network for solving a classification task using N-TIDIGITS18 dataset. This spike-based dataset consists of recordings from the Dynamic Audio Sensor, a spiking silicon cochlea sensor, in response to the TIDIGITS audio dataset. We also propose a new pre-processing method which applies an exponential kernel on the output cochlea spikes so that the interspike timing information is better preserved. The results from the N-TIDIGITS18 dataset show that the exponential features perform better than the spike count features, with over 91% accuracy on the digit classification task. This accuracy corresponds to an improvement of at least 2.5% over the use of spike count features, establishing a new state of the art for this dataset.

87 citations


Cites background or methods from "A 0.5 V 55 $\mu \text{W}$ 64 $\time..."

  • ..., 2015) and the Dynamic Audio Sensor (DAS) (Liu et al., 2014; Yang et al., 2016) fall roughly into two categories: either by the use of neural network methods or machine learning algorithms....

    [...]

  • ...INTRODUCTION The event processing methods for the asynchronous spikes of event-based sensors such as the Dynamic Vision Sensor (DVS) (Lichtsteiner et al., 2008; Berner et al., 2013; Posch et al., 2014; Yang et al., 2015) and the Dynamic Audio Sensor (DAS) (Liu et al., 2014; Yang et al., 2016) fall roughly into two categories: either by the use of neural network methods or machine learning algorithms....

    [...]

  • ...…Dynamic Vision Sensor (DVS) (Lichtsteiner et al., 2008; Berner et al., 2013; Posch et al., 2014; Yang et al., 2015) and the Dynamic Audio Sensor (DAS) (Liu et al., 2014; Yang et al., 2016) fall roughly into two categories: either by the use of neural network methods or machine learning algorithms....

    [...]

  • ...Time-binned SC features have been used for the speaker identification task using spike recordings generated from the TIMIT dataset (Liu et al., 2010; Li et al., 2012), the YOHO dataset (Chakrabartty and Liu, 2010), and real-world DAS recordings (Anumula et al., 2017)....

    [...]

  • ...The methods evaluated in this work were carried out on recordings from the CochleaAMS1b and CochleaAMS1c, while they will be evaluated on CochLP in the future....

    [...]

References
More filters
Journal ArticleDOI
08 Aug 2014-Science
TL;DR: Inspired by the brain’s structure, an efficient, scalable, and flexible non–von Neumann architecture is developed that leverages contemporary silicon technology and is well suited to many applications that use complex neural networks in real time, for example, multiobject detection and classification.
Abstract: Inspired by the brain’s structure, we have developed an efficient, scalable, and flexible non–von Neumann architecture that leverages contemporary silicon technology. To demonstrate, we built a 5.4-billion-transistor chip with 4096 neurosynaptic cores interconnected via an intrachip network that integrates 1 million programmable spiking neurons and 256 million configurable synapses. Chips can be tiled in two dimensions via an interchip communication interface, seamlessly scaling the architecture to a cortexlike sheet of arbitrary size. The architecture is well suited to many applications that use complex neural networks in real time, for example, multiobject detection and classification. With 400-pixel-by-240-pixel video input at 30 frames per second, the chip consumes 63 milliwatts.

3,253 citations


"A 0.5 V 55 $\mu \text{W}$ 64 $\time..." refers background in this paper

  • ...The benefits of this type of clockless system are threefold: 1) feature extraction in analog domain can save power, particularly when the required feature SNR is low to medium [16], [17]; 2) asynchronous event encoding helps reduce data redundancy, because the output event rate is proportionally dependent on the input activity and becomes zero if the input is quiescent [18], [19]; and 3) event-driven processors, such as continuous-time (CT) DSPs [19] and spiking neural networks [20], have adaptive power consumption that is correlated with the incoming event rate, and their processing latency can be reduced because of low data redundancy; low processing latency is particularly important for real-time applications involving human–machine interactions, where the machine’s immediate reaction to human expressions, such as speech, is often necessary....

    [...]

Book
01 Jan 1989
TL;DR: This chapter discusses a simple circuit that can generate a sinusoidal response and calls this circuit the second-order section, which can be used to generate any response that can be represented by two poles in the complex plane, where the two poles have both real and imaginary parts.

2,460 citations


"A 0.5 V 55 $\mu \text{W}$ 64 $\time..." refers background in this paper

  • ...6) with maximum 18 dB attenuation is used to extend the system upper bound input signal range while keeping the distortion of the following BPF low; inclusion of this attenuator is further motivated by the fact that active microphones can produce large amplitude output voltages with close-by loud sound even without standalone preamplifiers, which may cause large-signal oscillation of BPFs at high Q [43]....

    [...]

Journal ArticleDOI
22 Jan 2010
TL;DR: In this paper, the authors define and explore near-threshold computing (NTC), a design space where the supply voltage is approximately equal to the threshold voltage of the transistors.
Abstract: Power has become the primary design constraint for chip designers today. While Moore's law continues to provide additional transistors, power budgets have begun to prohibit those devices from actually being used. To reduce energy consumption, voltage scaling techniques have proved a popular technique with subthreshold design representing the endpoint of voltage scaling. Although it is extremely energy efficient, subthreshold design has been relegated to niche markets due to its major performance penalties. This paper defines and explores near-threshold computing (NTC), a design space where the supply voltage is approximately equal to the threshold voltage of the transistors. This region retains much of the energy savings of subthreshold operation with more favorable performance and variability characteristics. This makes it applicable to a broad range of power-constrained computing segments from sensors to high performance servers. This paper explores the barriers to the widespread adoption of NTC and describes current work aimed at overcoming these obstacles.

767 citations

01 Jan 2010
TL;DR: The barriers to the widespread adoption of near-threshold computing are explored and current work aimed at overcoming these obstacles are described.
Abstract: Power has become the primary design constraint for chip designers today. While Moore's law continues to provide additional transistors, power budgets have begun to prohibit those devices from actually being used. To reduce energy consumption, voltage scaling techniques have proved a popular technique with subthreshold design representing the endpoint of voltage scaling. Although it is extremely energy efficient, subthreshold design has been relegated to niche markets due to its major performance penalties. This paper defines and explores near-threshold computing (NTC), a design space where the supply voltage is approximately equal to the threshold voltage of the transistors. This region retains much of the energy savings of subthreshold operation with more favor- able performance and variability characteristics. This makes it applicable to a broad range of power-constrained computing segments from sensors to high performance servers. This paper explores the barriers to the widespread adoption of NTC and describes current work aimed at overcoming these obstacles.

695 citations


"A 0.5 V 55 $\mu \text{W}$ 64 $\time..." refers background in this paper

  • ...in the near-threshold region [56], [57] or completely avoided in an embedded system where the event streams are directly sent in parallel to the back-end event-driven processor on the same chip, as illustrated in Fig....

    [...]

Journal ArticleDOI
Rahul Sarpeshkar1
TL;DR: The results suggest that it is likely that the brain computes in a hybrid fashion and that an underappreciated and important reason for the efficiency of the human brain, which consumes only 12 W, is the hybrid and distributed nature of its architecture.
Abstract: We review the pros and cons of analog and digital computation. We propose that computation that is most efficient in its use of resources is neither analog computation nor digital computation but, rather, a mixture of the two forms. For maximum efficiency, the information and information-processing resources of the hybrid form must be distributed over many wires, with an optimal signal-to-noise ratio per wire. Our results suggest that it is likely that the brain computes in a hybrid fashion and that an underappreciated and important reason for the efficiency of the human brain, which consumes only 12 W, is the hybrid and distributed nature of its architecture.

495 citations


"A 0.5 V 55 $\mu \text{W}$ 64 $\time..." refers background in this paper

  • ...The benefits of this type of clockless system are threefold: 1) feature extraction in analog domain can save power, particularly when the required feature SNR is low to medium [16], [17]; 2) asynchronous event encoding helps reduce data redundancy, because the output event rate is proportionally dependent on the input activity and becomes zero if the input is quiescent [18], [19]; and 3) event-driven processors, such as continuous-time (CT) DSPs [19] and spiking neural networks [20], have adaptive power consumption that is correlated with the incoming event rate, and their processing latency can be reduced because of low data redundancy; low processing latency is particularly important for real-time applications involving human–machine interactions, where the machine’s immediate reaction to human expressions, such as speech, is often necessary....

    [...]