scispace - formally typeset
Search or ask a question
Topic

Microphone

About: Microphone is a research topic. Over the lifetime, 39999 publications have been published within this topic receiving 337352 citations. The topic is also known as: mic & mike.


Papers
More filters
Journal ArticleDOI
TL;DR: In this article, four different approaches are used to determine experimentally the sources of jet mixing noise: spectral and directional information measured by a single microphone in the far field, fine-scale turbulence, large turbulence structures of the jet flow, and a mirror microphone is used to measure the noise source distribution along the lengths of high speed jets.
Abstract: The primary objective of this investigation is to determine experimentally the sources of jet mixing noise. In the present study, four different approaches are used. It is reasonable to assume that the characteristics of the noise sources are imprinted on their radiation fields. Under this assumption, it becomes possible to analyse the characteristics of the far-field sound and then infer back to the characteristics of the sources. The first approach is to make use of the spectral and directional information measured by a single microphone in the far field. A detailed analysis of a large collection of far-field noise data has been carried out. The purpose is to identify special characteristics that can be linked directly to those of the sources. The second approach is to measure the coherence of the sound field using two microphones. The autocorrelations and cross-correlations of these measurements offer not only valuable information on the spatial structure of the noise field in the radial and polar angle directions, but also on the sources inside the jet. The third approach involves measuring the correlation between turbulence fluctuations inside a jet and the radiated noise in the far field. This is the most direct and unambiguous way of identifying the sources of jet noise. In the fourth approach, a mirror microphone is used to measure the noise source distribution along the lengths of high-speed jets. Features and trends observed in noise source strength distributions are expected to shed light on the source mechanisms. It will be shown that all four types of data indicate clearly the existence of two distinct noise sources in jets. One source of noise is the fine-scale turbulence and the other source is the large turbulence structures of the jet flow. Some of the salient features of the sound field associated with the two noise sources are reported in this paper.

486 citations

Journal ArticleDOI
TL;DR: In this article, the authors describe a transducer system that discriminates against sound arrivals from all directions except for that of the desired source, using a two-dimensional array of microphones.
Abstract: The quality of sound pickup in large rooms—such as auditoria, conference rooms, or classrooms—is impaired by reverberation and interfering noise sources. These degradations can be minimized by a transducer system that discriminates against sound arrivals from all directions except for that of the desired source. A two‐dimensional array of microphones can be electronically beam steered to accomplish this directivity. This report gives the theory, design, and implementation of a microprocessor system for automatically steering a two‐dimensional microphone array. The signal‐seeking transducer system is implemented as a dual‐beam, “track‐while‐scan” array. It utilizes signal properties to distinguish between desired speech sources and interfering noise. The complete automatic system has been tested in anechoic and medium‐sized auditorium environments, and its performance is discussed.

484 citations

Proceedings Article
01 Jan 2000
TL;DR: A technique called refiltering is presented which recovers sources by a nonstationary reweighting of frequency sub-bands from a single recording, and it is argued for the application of statistical algorithms to learning this masking function.
Abstract: Source separation, or computational auditory scene analysis, attempts to extract individual acoustic objects from input which contains a mixture of sounds from different sources, altered by the acoustic environment. Unmixing algorithms such as ICA and its extensions recover sources by reweighting multiple observation sequences, and thus cannot operate when only a single observation signal is available. I present a technique called refiltering which recovers sources by a nonstationary reweighting ("masking") of frequency sub-bands from a single recording, and argue for the application of statistical algorithms to learning this masking function. I present results of a simple factorial HMM system which learns on recordings of single speakers and can then separate mixtures using only one observation signal by computing the masking function and then refiltering.

476 citations

BookDOI
01 May 1991
TL;DR: This dissertation describes a number of algorithms developed to increase the robustness of automatic speech recognition systems with respect to changes in the environment, including the SNR-Dependent Cepstral Normalization, (SDCN) and the Codeword-Dependent Cep stral normalization (CDCN).
Abstract: This dissertation describes a number of algorithms developed to increase the robustness of automatic speech recognition systems with respect to changes in the environment. These algorithms attempt to improve the recognition accuracy of speech recognition systems when they are trained and tested in different acoustical environments, and when a desk-top microphone (rather than a close-talking microphone) is used for speech input. Without such processing, mismatches between training and testing conditions produce an unacceptable degradation in recognition accuracy. Two kinds of environmental variability are introduced by the use of desk-top microphones and different training and testing conditions: additive noise and spectral tilt introduced by linear filtering. An important attribute of the novel compensation algorithms described in this thesis is that they provide joint rather than independent compensation for these two types of degradation. Acoustical compensation is applied in our algorithms as an additive correction in the cepstral domain. This allows a higher degree of integration within SPHINX, the Carnegie Mellon speech recognition system, that uses the cepstrum as its feature vector. Therefore, these algorithms can be implemented very efficiently. Processing in many of these algorithms is based on instantaneous signal-to-noise ratio (SNR), as the appropriate compensation represents a form of noise suppression at low SNRs and spectral equalization at high SNRs. The compensation vectors for additive noise and spectral transformations are estimated by minimizing the differences between speech feature vectors obtained from a "standard" training corpus of speech and feature vectors that represent the current acoustical environment. In our work this is accomplished by minimizing the distortion of vector-quantized cepstra that are produced by the feature extraction module in SPHINX. In this dissertation we describe several algorithms including the SNR-Dependent Cepstral Normalization, (SDCN) and the Codeword-Dependent Cepstral Normalization (CDCN). With CDCN, the accuracy of SPHINX when trained on speech recorded with a close-talking microphone and tested on speech recorded with a desk-top microphone is essentially the same obtained when the system is trained and tested on speech from the desk-top microphone. An algorithm for frequency normalization has also been proposed in which the parameter of the bilinear transformation that is used by the signal-processing stage to produce frequency warping is adjusted for each new speaker and acoustical environment. The optimum value of this parameter is again chosen to minimize the vector-quantization distortion between the standard environment and the current one. In preliminary studies, use of this frequency normalization produced a moderate additional decrease in the observed error rate.

474 citations

Proceedings ArticleDOI
03 Apr 1990
TL;DR: Initial efforts to make Sphinx, a continuous-speech speaker-independent recognition system, robust to changes in the environment are reported, and two novel methods based on additive corrections in the cepstral domain are proposed.
Abstract: Initial efforts to make Sphinx, a continuous-speech speaker-independent recognition system, robust to changes in the environment are reported. To deal with differences in noise level and spectral tilt between close-talking and desk-top microphones, two novel methods based on additive corrections in the cepstral domain are proposed. In the first algorithm, the additive correction depends on the instantaneous SNR of the signal. In the second technique, expectation-maximization techniques are used to best match the cepstral vectors of the input utterances to the ensemble of codebook entries representing a standard acoustical ambience. Use of the algorithms dramatically improves recognition accuracy when the system is tested on a microphone other than the one on which it was trained. >

461 citations


Network Information
Related Topics (5)
Speech processing
24.2K papers, 637K citations
86% related
Noise
110.4K papers, 1.3M citations
82% related
Transducer
58.1K papers, 681.8K citations
80% related
Signal processing
73.4K papers, 983.5K citations
79% related
Frequency domain
53.8K papers, 701.3K citations
78% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
2023493
2022985
2021670
20201,638
20191,955
20182,056