scispace - formally typeset
Open AccessJournal ArticleDOI

Molecular Property Prediction Based on a Multichannel Substructure Graph

TLDR
A multichannel substructure-graph gated recurrent unit (GRU) architecture is proposed, which is a novel GRU-based neural network with attention mechanisms applied to molecular substructures to learn and predict properties.
Abstract
Molecular property prediction is important to drug design. With the development of artificial intelligence, deep learning methods are effective for extracting molecular features. In this paper, we propose a multichannel substructure-graph gated recurrent unit (GRU) architecture, which is a novel GRU-based neural network with attention mechanisms applied to molecular substructures to learn and predict properties. In the architecture, molecular features are extracted at the node level and molecule level for capturing fine-grained and coarse-grained information. In addition, three bidirectional GRUs are adopted to extract the features on three channels to generate the molecular representations. Different attention weights are assigned to the entities in the molecule to evaluate their contributions. Experiments are implemented to compare our model with benchmark models in molecular property prediction for both regression and classification tasks, and the results show that our model has strong robustness and generalizability.

read more

Content maybe subject to copyright    Report

Citations
More filters

Convolutional Embedding of Attributed Molecular Graphs for Physical Property Prediction

TL;DR: In this article, simple atom and bond attributes are used to construct atom-specific feature vectors that take into account the local chemical environment using different neighborhood radii, which significantly enhances model performance.
Journal ArticleDOI

Gas Chromatographic Retention Index Prediction Using Multimodal Machine Learning

TL;DR: In this article, four machine learning models for retention index prediction: 1D and 2D convolutional neural networks, deep residual multilayer perceptron, and gradient boosting.
Journal ArticleDOI

MCN-CPI: Multiscale Convolutional Network for Compound-Protein Interaction Prediction.

TL;DR: Wang et al. as discussed by the authors proposed a multiscale convolutional network that extracted the local and global features of the protein and the topological feature of the compound using different types of CNNs.
Journal ArticleDOI

Different molecular enumeration influences in deep learning: an example using aqueous solubility.

TL;DR: This work carefully examined the commonly used simplified molecular-input line-entry specification (SMILES) notation representing a single molecule and proposed to use the full enumerations in SMILES to achieve better accuracy and demonstrated the use of attention in the decoding network to detect the part of a molecule that is relevant to solubility.
Journal ArticleDOI

Molecular substructure tree generative model for de novo drug design

TL;DR: The proposed model is based on the Variational Auto-Encoder architecture, which uses the encoder to map molecules to the latent vector space, and then builds an autoregressive generative model as a decoder to generate new molecules from Gaussian distribution.
References
More filters
Proceedings Article

Adam: A Method for Stochastic Optimization

TL;DR: This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Journal ArticleDOI

Support-Vector Networks

TL;DR: High generalization ability of support-vector networks utilizing polynomial input transformations is demonstrated and the performance of the support- vector network is compared to various classical learning algorithms that all took part in a benchmark study of Optical Character Recognition.
Journal ArticleDOI

Greedy function approximation: A gradient boosting machine.

TL;DR: A general gradient descent boosting paradigm is developed for additive expansions based on any fitting criterion, and specific algorithms are presented for least-squares, least absolute deviation, and Huber-M loss functions for regression, and multiclass logistic likelihood for classification.
Posted Content

Semi-Supervised Classification with Graph Convolutional Networks

TL;DR: A scalable approach for semi-supervised learning on graph-structured data that is based on an efficient variant of convolutional neural networks which operate directly on graphs which outperforms related methods by a significant margin.
Posted Content

Sequence to Sequence Learning with Neural Networks

TL;DR: This paper presents a general end-to-end approach to sequence learning that makes minimal assumptions on the sequence structure, and finds that reversing the order of the words in all source sentences improved the LSTM's performance markedly, because doing so introduced many short term dependencies between the source and the target sentence which made the optimization problem easier.
Related Papers (5)