scispace - formally typeset
Open AccessProceedings Article

Framewise phoneme classification with bidirectional LSTM and other neural network architectures

Alex Graves, +1 more
- Vol. 18, pp 602-610
Reads0
Chats0
TLDR
In this article, a modified, full gradient version of the LSTM learning algorithm was used for framewise phoneme classification, using the TIMIT database, and the results support the view that contextual information is crucial to speech processing, and suggest that bidirectional networks outperform unidirectional ones.
Abstract
In this paper, we present bidirectional Long Short Term Memory (LSTM) networks, and a modified, full gradient version of the LSTM learning algorithm. We evaluate Bidirectional LSTM (BLSTM) and several other network architectures on the benchmark task of framewise phoneme classification, using the TIMIT database. Our main findings are that bidirectional networks outperform unidirectional ones, and Long Short Term Memory (LSTM) is much faster and also more accurate than both standard Recurrent Neural Nets (RNNs) and time-windowed Multilayer Perceptrons (MLPs). Our results support the view that contextual information is crucial to speech processing, and suggest that BLSTM is an effective architecture with which to exploit it'.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Spectral features based convolutional neural network for accurate and prompt identification of schizophrenic patients

TL;DR: The overall analysis of simulation results of all models exhibits that the proposed spectral features based CNN model is an efficient technique for accurate and prompt identification of schizophrenic patients among healthy individuals with average classification accuracies of 94.08% and 98.56% for two different datasets with optimally small classification time.
Proceedings ArticleDOI

Here's My Point: Joint Pointer Architecture for Argument Mining.

TL;DR: The authors proposed a neural network-based approach to link extraction in argument mining, which achieved state-of-the-art results on two separate evaluation corpora, showing far superior performance than the previously proposed corpus-specific and heavily feature-engineered models.

Collaborative Learning on the Edges: A Case Study on Connected Vehicles

TL;DR: This work proposes CLONE, a collaborative learning setting on the edges based on the real-world dataset collected from a large electric vehicle (EV) company, built on top of the federated learning algorithm and long shortterm memory networks, and demonstrates the effectiveness of driver personalization, privacy serving, latency reduction (asynchronous execution), and security protection.
Journal ArticleDOI

Utilizing Student Time Series Behaviour in Learning Management Systems for Early Prediction of Course Performance

TL;DR: Results indicate that using thedeep learning approach, time series information about click frequencies successfully provided early detection of at-risk students with moderate prediction accuracy and the deep learning approach showed higher prediction performance and stronger generalizability than the machine learning classifiers.
Posted Content

Here's My Point: Joint Pointer Architecture for Argument Mining

TL;DR: This work presents the first neural network-based approach to link extraction in argument mining, and proposes a novel architecture that applies Pointer Network sequence-to-sequence attention modeling to structural prediction in discourse parsing tasks and develops a joint model that achieves state-of-the-art results.
References
More filters
Journal ArticleDOI

Long short-term memory

TL;DR: A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.
Book

Neural networks for pattern recognition

TL;DR: This is the first comprehensive treatment of feed-forward neural networks from the perspective of statistical pattern recognition, and is designed as a text, with over 100 exercises, to benefit anyone involved in the fields of neural computation and pattern recognition.
Journal ArticleDOI

Bidirectional recurrent neural networks

TL;DR: It is shown how the proposed bidirectional structure can be easily modified to allow efficient estimation of the conditional posterior probability of complete symbol sequences without making any explicit assumption about the shape of the distribution.

Gradient Flow in Recurrent Nets: the Difficulty of Learning Long-Term Dependencies

TL;DR: D3EGF(FIH)J KMLONPEGQSRPETN UCV.
Related Papers (5)