scispace - formally typeset
Open AccessJournal ArticleDOI

Artificial Neural Networks in Action for an Automated Cell-Type Classification of Biological Neural Networks

TLDR
This work examines three different deep learning models aiming at an automated neuronal cell-type classification and compare their performance, and demonstrates the efficacy and potent capabilities for each one of the proposed schemes.
Abstract
Identification of different neuronal cell types is critical for understanding their contribution to brain functions. Yet, automated and reliable classification of neurons remains a challenge, primarily because of their biological complexity. Typical approaches include laborious and expensive immunohistochemical analysis while feature extraction algorithms based on cellular characteristics have recently been proposed. The former rely on molecular markers, which are often expressed in many cell types, while the latter suffer from similar issues: finding features that are distinctive for each class has proven to be equally challenging. Moreover, both approaches are time consuming and demand a lot of human intervention. In this work we establish the first, automated cell-type classification method that relies on neuronal activity rather than molecular or cellular features. We test our method on a real-world dataset comprising of raw calcium activity signals for four neuronal types. We compare the performance of three different deep learning models and demonstrate that our method can achieve automated classification of neuronal cell types with unprecedented accuracy.

read more

Citations
More filters
Journal ArticleDOI

Solving multi-objective optimization problem of convolutional neural network using fast forward quantum optimization algorithm: Application in digital image classification

TL;DR: The FFQOAconNetwork as discussed by the authors is a hybridization of fast forward quantum optimization (FFQOA) with CNN, which searches for the optimal weights associated with the layers by simultaneously achieving the minimal classification error.
Journal ArticleDOI

Cortical Pyramidal and Parvalbumin Cells Exhibit Distinct Spatiotemporal Extracellular Electric Potentials

- 01 Nov 2022 - 
TL;DR: In this paper , the spatial distribution of extracellular voltage during a spike depends on cellular morphology, connectivity, and identity, and the success of binary classification of pyramidal cells and parvalbumin-immunoreactive (PV) cells was demonstrated.
Journal ArticleDOI

Cortical Pyramidal and Parvalbumin Cells Exhibit Distinct Spatiotemporal Extracellular Electric Potentials

Lior J Sukman, +1 more
- 17 Nov 2022 - 
TL;DR: In this article , the spatial distribution of extracellular voltage during a spike depends on cellular morphology, connectivity, and identity, and the success of binary classification of pyramidal cells and parvalbumin-immunoreactive (PV) cells was demonstrated.
Journal ArticleDOI

Intra Prediction Method for Depth Video Coding by Block Clustering through Deep Learning

Dong-Seok Lee, +1 more
- 01 Dec 2022 - 
TL;DR: In this paper , the authors proposed an intra-picture prediction method for depth video by a block clustering through a neural network consisting of both a spatial feature prediction network and a clustering network.
Journal ArticleDOI

A Generative Neighborhood-based Deep Autoencoder for Robust Imbalanced Classification

TL;DR: GenDA as discussed by the authors is a generative neighborhood-based deep autoencoder, which can be successfully applied to both image and time series data, and can be used to learn latent representations based on the neighboring embedding space of the samples.
References
More filters
Journal ArticleDOI

Long short-term memory

TL;DR: A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.
Journal Article

Dropout: a simple way to prevent neural networks from overfitting

TL;DR: It is shown that dropout improves the performance of neural networks on supervised learning tasks in vision, speech recognition, document classification and computational biology, obtaining state-of-the-art results on many benchmark data sets.
Journal ArticleDOI

Approximation capabilities of multilayer feedforward networks

TL;DR: It is shown that standard multilayer feedforward networks with as few as a single hidden layer and arbitrary bounded and nonconstant activation function are universal approximators with respect to L p (μ) performance criteria, for arbitrary finite input environment measures μ.
Proceedings Article

Framewise phoneme classification with bidirectional LSTM and other neural network architectures

TL;DR: In this article, a modified, full gradient version of the LSTM learning algorithm was used for framewise phoneme classification, using the TIMIT database, and the results support the view that contextual information is crucial to speech processing, and suggest that bidirectional networks outperform unidirectional ones.
Related Papers (5)