scispace - formally typeset
Proceedings ArticleDOI

Identification of Audio Processing Operations Based on Convolutional Neural Network

Reads0
Chats0
TLDR
The experimental results show that the proposed convolutional neural network to detect audio processing operations can significantly outperform related methods based on hand-crafted features and other CNN architectures, and can achieve state-of-the-art results for both binary and multiple classification.
Abstract
To reduce the tampering artifacts and/or enhance audio quality, some audio processing operations are often applied in the resulting tampered audio. Like image forensics, the detection of various post processing operations has become very important for audio authentication. In this paper, we propose a convolutional neural network (CNN) to detect audio processing operations. In the proposed method, we carefully design the network architecture, with particular attention to the frequency representation for the audio input, the activation function and the depth of the network. In our experiments, we evaluate the proposed method on audio clips with 12 commonly used audio processing operations and of three different small sizes. The experimental results show that our method can significantly outperform related methods based on hand-crafted features and other CNN architectures, and can achieve state-of-the-art results for both binary and multiple classification.

read more

Citations
More filters
Journal ArticleDOI

Identification of Weakly Pitch-Shifted Voice Based on Convolutional Neural Network

TL;DR: A convolutional neural network is proposed to detect not only strongly pitch-shifted voice but also weakly pitch- shifted voice of which the shifting factor is less than ±4 semitones.
Proceedings ArticleDOI

How Initialization is Related to Deep Neural Networks Generalization Capability: Experimental Study

TL;DR: The focus of this paper is on improving the generalization ability, which is a key for successful implementation of deep neural networks, and an experimental study is done to answer the question how initialization is related to theDeep neural networks' generalization capability.
Book ChapterDOI

Detection of Various Speech Forgery Operations Based on Recurrent Neural Network

TL;DR: In this paper, a forensic algorithm based on recurrent neural network (RNN) and linear frequency cepstrum coefficients (LFCC) is proposed to detect four common forgery operations.
Journal ArticleDOI

Efficacy of Residual Methods for Passive Image Forensics Using Four Filtered Residue CNN

TL;DR: The generalization ability and high detection accuracy in the presence of anti-forensics operation highlight the efficacy of the proposed FFR-CNN, and constrained time complexity supports the effectiveness of FFR–CNN for real time applications.
References
More filters
Journal Article

Dropout: a simple way to prevent neural networks from overfitting

TL;DR: It is shown that dropout improves the performance of neural networks on supervised learning tasks in vision, speech recognition, document classification and computational biology, obtaining state-of-the-art results on many benchmark data sets.
Posted Content

Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification

TL;DR: This work proposes a Parametric Rectified Linear Unit (PReLU) that generalizes the traditional rectified unit and derives a robust initialization method that particularly considers the rectifier nonlinearities.
Proceedings ArticleDOI

Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification

TL;DR: In this paper, a Parametric Rectified Linear Unit (PReLU) was proposed to improve model fitting with nearly zero extra computational cost and little overfitting risk, which achieved a 4.94% top-5 test error on ImageNet 2012 classification dataset.
Proceedings Article

On the importance of initialization and momentum in deep learning

TL;DR: It is shown that when stochastic gradient descent with momentum uses a well-designed random initialization and a particular type of slowly increasing schedule for the momentum parameter, it can train both DNNs and RNNs to levels of performance that were previously achievable only with Hessian-Free optimization.
Related Papers (5)