Molecular Property Prediction Based on a Multichannel Substructure Graph
TLDR
A multichannel substructure-graph gated recurrent unit (GRU) architecture is proposed, which is a novel GRU-based neural network with attention mechanisms applied to molecular substructures to learn and predict properties.Abstract:
Molecular property prediction is important to drug design. With the development of artificial intelligence, deep learning methods are effective for extracting molecular features. In this paper, we propose a multichannel substructure-graph gated recurrent unit (GRU) architecture, which is a novel GRU-based neural network with attention mechanisms applied to molecular substructures to learn and predict properties. In the architecture, molecular features are extracted at the node level and molecule level for capturing fine-grained and coarse-grained information. In addition, three bidirectional GRUs are adopted to extract the features on three channels to generate the molecular representations. Different attention weights are assigned to the entities in the molecule to evaluate their contributions. Experiments are implemented to compare our model with benchmark models in molecular property prediction for both regression and classification tasks, and the results show that our model has strong robustness and generalizability.read more
Citations
More filters
Convolutional Embedding of Attributed Molecular Graphs for Physical Property Prediction
TL;DR: In this article, simple atom and bond attributes are used to construct atom-specific feature vectors that take into account the local chemical environment using different neighborhood radii, which significantly enhances model performance.
Journal ArticleDOI
Gas Chromatographic Retention Index Prediction Using Multimodal Machine Learning
TL;DR: In this article, four machine learning models for retention index prediction: 1D and 2D convolutional neural networks, deep residual multilayer perceptron, and gradient boosting.
Journal ArticleDOI
MCN-CPI: Multiscale Convolutional Network for Compound-Protein Interaction Prediction.
TL;DR: Wang et al. as discussed by the authors proposed a multiscale convolutional network that extracted the local and global features of the protein and the topological feature of the compound using different types of CNNs.
Journal ArticleDOI
Different molecular enumeration influences in deep learning: an example using aqueous solubility.
Jen-Hao Chen,Yufeng J. Tseng +1 more
TL;DR: This work carefully examined the commonly used simplified molecular-input line-entry specification (SMILES) notation representing a single molecule and proposed to use the full enumerations in SMILES to achieve better accuracy and demonstrated the use of attention in the decoding network to detect the part of a molecule that is relevant to solubility.
Journal ArticleDOI
Molecular substructure tree generative model for de novo drug design
TL;DR: The proposed model is based on the Variational Auto-Encoder architecture, which uses the encoder to map molecules to the latent vector space, and then builds an autoregressive generative model as a decoder to generate new molecules from Gaussian distribution.
References
More filters
Proceedings Article
Adam: A Method for Stochastic Optimization
Diederik P. Kingma,Jimmy Ba +1 more
TL;DR: This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Journal ArticleDOI
Support-Vector Networks
Corinna Cortes,Vladimir Vapnik +1 more
TL;DR: High generalization ability of support-vector networks utilizing polynomial input transformations is demonstrated and the performance of the support- vector network is compared to various classical learning algorithms that all took part in a benchmark study of Optical Character Recognition.
Journal ArticleDOI
Greedy function approximation: A gradient boosting machine.
TL;DR: A general gradient descent boosting paradigm is developed for additive expansions based on any fitting criterion, and specific algorithms are presented for least-squares, least absolute deviation, and Huber-M loss functions for regression, and multiclass logistic likelihood for classification.
Posted Content
Semi-Supervised Classification with Graph Convolutional Networks
Thomas Kipf,Max Welling +1 more
TL;DR: A scalable approach for semi-supervised learning on graph-structured data that is based on an efficient variant of convolutional neural networks which operate directly on graphs which outperforms related methods by a significant margin.
Posted Content
Sequence to Sequence Learning with Neural Networks
TL;DR: This paper presents a general end-to-end approach to sequence learning that makes minimal assumptions on the sequence structure, and finds that reversing the order of the words in all source sentences improved the LSTM's performance markedly, because doing so introduced many short term dependencies between the source and the target sentence which made the optimization problem easier.
Related Papers (5)
Molecular Mechanics-Driven Graph Neural Network with Multiplex Graph for Molecular Structures.
Shuo Zhang,Yang Liu,Lei Xie +2 more
Heterogeneous Molecular Graph Neural Networks for Predicting Molecule Properties
Zeren Shui,George Karypis +1 more
A chemical group graph representation for efficient high-throughput analysis of atomistic protein simulations
Noah C. Benson,Valerie Daggett +1 more