scispace - formally typeset
Open AccessProceedings Article

Multimodal Deep Learning

Reads0
Chats0
TLDR
This work presents a series of tasks for multimodal learning and shows how to train deep networks that learn features to address these tasks, and demonstrates cross modality feature learning, where better features for one modality can be learned if multiple modalities are present at feature learning time.
Abstract
Deep networks have been successfully applied to unsupervised feature learning for single modalities (e.g., text, images or audio). In this work, we propose a novel application of deep networks to learn features over multiple modalities. We present a series of tasks for multimodal learning and show how to train deep networks that learn features to address these tasks. In particular, we demonstrate cross modality feature learning, where better features for one modality (e.g., video) can be learned if multiple modalities (e.g., audio and video) are present at feature learning time. Furthermore, we show how to learn a shared representation between modalities and evaluate it on a unique task, where the classifier is trained with audio-only data but tested with video-only data and vice-versa. Our models are validated on the CUAVE and AVLetters datasets on audio-visual speech classification, demonstrating best published visual speech classification on AVLetters and effective shared representation learning.

read more

Content maybe subject to copyright    Report

Citations
More filters
DissertationDOI

Geometry and Uncertainty in Deep Learning for Computer Vision

Alex Kendall
TL;DR: This thesis presents end-to-end deep learning architectures for a number of core computer vision problems; scene understanding, camera pose estimation, stereo vision and video semantic segmentation, and introduces ideas from probabilistic modelling and Bayesian deep learning to understand uncertainty in computer vision models.
Proceedings ArticleDOI

Towards 3D object detection with bimodal deep Boltzmann machines over RGBD imagery

TL;DR: This work proposes a cross-modality deep learning framework based on deep Boltzmann Machines for 3D Scenes object detection and demonstrates that by learning cross- modality feature from RGBD data, it is possible to capture their joint information to reinforce detector trainings in individual modalities.
Posted Content

Y^2Seq2Seq: Cross-Modal Representation Learning for 3D Shape and Text by Joint Reconstruction and Prediction of View and Word Sequences

TL;DR: This work proposes Y^2Seq2Sequ, a view-based model, to learn cross-modal representations by joint reconstruction and prediction of view and word sequences, and bridges the semantic meaning embedded in the two modalities by two coupled `Y' like sequence-to-sequence structures.
Journal ArticleDOI

SPARCNet: A Hardware Accelerator for Efficient Deployment of Sparse Convolutional Networks

TL;DR: The proposed SPARCNet, a hardware accelerator for efficient deployment of SPARse Convolutional NETworks, looks to enable deploying networks in embedded, resource-bound settings by both exploiting efficient forms of parallelism inherent in convolutional layers and by exploiting the sparsification and approximation techniques proposed.
Posted Content

Deep Partial Multi-View Learning.

TL;DR: Cross Partial Multi-View Networks (CPM-Nets) as discussed by the authors is a framework for multi-view representation learning, which aims to fully and flflexibly take advantage of multiple partial views.
References
More filters
Proceedings ArticleDOI

Histograms of oriented gradients for human detection

TL;DR: It is shown experimentally that grids of histograms of oriented gradient (HOG) descriptors significantly outperform existing feature sets for human detection, and the influence of each stage of the computation on performance is studied.
Journal ArticleDOI

Reducing the Dimensionality of Data with Neural Networks

TL;DR: In this article, an effective way of initializing the weights that allows deep autoencoder networks to learn low-dimensional codes that work much better than principal components analysis as a tool to reduce the dimensionality of data is described.
Journal ArticleDOI

A fast learning algorithm for deep belief nets

TL;DR: A fast, greedy algorithm is derived that can learn deep, directed belief networks one layer at a time, provided the top two layers form an undirected associative memory.
Proceedings ArticleDOI

Extracting and composing robust features with denoising autoencoders

TL;DR: This work introduces and motivate a new training principle for unsupervised learning of a representation based on the idea of making the learned representations robust to partial corruption of the input pattern.
Journal ArticleDOI

Hearing lips and seeing voices

TL;DR: The study reported here demonstrates a previously unrecognised influence of vision upon speech perception, on being shown a film of a young woman's talking head in which repeated utterances of the syllable [ba] had been dubbed on to lip movements for [ga].
Related Papers (5)