scispace - formally typeset
Search or ask a question
Topic

Audio signal processing

About: Audio signal processing is a research topic. Over the lifetime, 21463 publications have been published within this topic receiving 319597 citations. The topic is also known as: audio processing & Acoustic signal processing.


Papers
More filters
Proceedings ArticleDOI
21 Oct 2001
TL;DR: Preliminary experimental results suggest that the listener's ability to identify messages in a multi-talker environment significantly improves by enhancing a monophonic signal with the proposed scheme.
Abstract: We introduce a new scheme for simultaneous placement of a number of sources in auditory space. The scheme is based on an assumption about the relevance of localization cues in different critical bands. Given the sum signal of a number of sources, i.e. a monophonic signal, and a set of parameters (side-information) the scheme is capable of generating a binaural signal by spatially placing the sources contained in the monophonic signal. Potential applications for the scheme are multi-talker desktop conferencing and audio coding. Preliminary experimental results suggest that the listener's ability to identify messages in a multi-talker environment significantly improves by enhancing a monophonic signal with the proposed scheme.

115 citations

Patent
28 Feb 1992
TL;DR: In this article, a global positioning system receiver method and system which converts the analog signals to digital signals prior to performinng signal acquisition and GPS unique processing functions is disclosed, which allows full multi channel tracking with several channels (202, 203, 204) each tracking a separate signal while the remainder of the channels rove over all of the signals on the alternate L-band with programmable duty cycles.
Abstract: There is disclosed a global positioning system receiver method and system which converts the analog signals to digital signals prior to performinng signal acquisition and GPS unique processing functions. The A/D converter (201) uses full null zone processing to increase anti-jamming capability and 4-level output to reduce signal processing. A. single analog and digital channel (302) is used for both the L1 and L2 channels as well as for all P(Y)-code and C/A-code thus allowing full multi channel tracking with several channels (202, 203, 204) each tracking a separate signal while the remainder of the channels rove over all of the signals on the alternate L-band with programmable duty cycles. The system allows for Y-code substitution for P-code by multiplexing a portion of the Y-code generated between multiple channels.

114 citations

Proceedings ArticleDOI
01 Jun 2019
TL;DR: A deep convolutional neural network is devised that learns to decode the monaural soundtrack into its binaural counterpart by injecting visual information about object and scene configurations, and the resulting output 2.5D visual sound helps "lift" the flat single channel audio into spatialized sound.
Abstract: Binaural audio provides a listener with 3D sound sensation, allowing a rich perceptual experience of the scene. However, binaural recordings are scarcely available and require nontrivial expertise and equipment to obtain. We propose to convert common monaural audio into binaural audio by leveraging video. The key idea is that visual frames reveal significant spatial cues that, while explicitly lacking in the accompanying single-channel audio, are strongly linked to it. Our multi-modal approach recovers this link from unlabeled video. We devise a deep convolutional neural network that learns to decode the monaural (single-channel) soundtrack into its binaural counterpart by injecting visual information about object and scene configurations. We call the resulting output 2.5D visual sound---the visual stream helps "lift" the flat single channel audio into spatialized sound. In addition to sound generation, we show the self-supervised representation learned by our network benefits audio-visual source separation. Our video results: http://vision.cs.utexas.edu/projects/2.5D_visual_sound/

114 citations

PatentDOI
TL;DR: In this paper, a kinetic device actuating signal, sampling keyboard based encoder is coupled via an audio mixer to an audiovisual programming source and television transmitter carrying an audio based kinetic and audio signal complex.
Abstract: Movable and audible toys and other animated devices (14) spaced apart from a television screen are provided with program synchronized audio and control data to interreact with the program viewer in relationship to its programming. A kinetic device actuating signal, sampling keyboard based encoder (12) is coupled via an audio mixer to an audiovisual programming source and television transmitter (16) carrying an audio based kinetic and audio signal complex. At a remote location (200), coded audio and kinetic device signals along with audiovisual programming are received and the audiovisual programming content is displayed for viewing and listening. Stereo sound band based program audio signals are decoded and separated from the stereo sound band based device audio and kinetic signals. The device audio and kinetic signals are retransmitted to a spaced-apart toy causing the device to be audible and to move in synchronization with the spaced-apart audiovisual programming. A single band low powered FM transmitter (28) carries the audio kinetic signal complex in proximity yet spaced apart from a receiver coupled to an audio cassette configured magnetic head transducer disposed in the kinetic device to communicate audio and kinetic information to a device speaker and motors to cause the device to move and be audible.

114 citations

Patent
12 Mar 2014
TL;DR: In this article, a variation of a method for augmenting a listening experience of a user through an audio device includes detecting a location of the audio device; selecting a set of audio output feedbacks; identifying a common feature across audio outputs; transforming an audio signal into a processed audio signal according to a hearing profile of the user and the common feature.
Abstract: One variation of a method for augmenting a listening experience of a user through an audio device includes: detecting a location of the audio device; selecting a set of audio output feedbacks, each audio output feedback in the set of audio output feedback entered by an individual and associated with a physical site proximal to the location; identifying a common feature across audio output feedbacks within the set of audio output feedbacks; transforming an audio signal into a processed audio signal according to a hearing profile of the user and the common feature; and outputting the processed audio signal through the audio device.

114 citations


Network Information
Related Topics (5)
Feature extraction
111.8K papers, 2.1M citations
81% related
Feature (computer vision)
128.2K papers, 1.7M citations
79% related
Robustness (computer science)
94.7K papers, 1.6M citations
78% related
Noise
110.4K papers, 1.3M citations
77% related
Image segmentation
79.6K papers, 1.8M citations
77% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
202319
202263
2021217
2020525
2019659
2018597