A survey on graph kernels
Reads0
Chats0
TLDR
This survey gives a comprehensive overview of techniques for kernel-based graph classification developed in the past 15 years and describes and categorizes graph kernels based on properties inherent to their design, such as the nature of their extracted graph features, their method of computation and their applicability to problems in practice.Abstract:
Graph kernels have become an established and widely-used technique for solving classification tasks on graphs. This survey gives a comprehensive overview of techniques for kernel-based graph classification developed in the past 15 years. We describe and categorize graph kernels based on properties inherent to their design, such as the nature of their extracted graph features, their method of computation and their applicability to problems in practice. In an extensive experimental evaluation, we study the classification accuracy of a large suite of graph kernels on established benchmarks as well as new datasets. We compare the performance of popular kernels with several baseline methods and study the effect of applying a Gaussian RBF kernel to the metric induced by a graph kernel. In doing so, we find that simple baselines become competitive after this transformation on some datasets. Moreover, we study the extent to which existing graph kernels agree in their predictions (and prediction errors) and obtain a data-driven categorization of kernels as result. Finally, based on our experimental results, we derive a practitioner’s guide to kernel-based graph classification.read more
Citations
More filters
Book
Graph Representation Learning
TL;DR: This work has shown that graph-structured data is ubiquitous throughout the natural and social sciences, from telecommunication networks to quantum chemistry, and building relational inductive biases into deep learnin...
Posted Content
TUDataset: A collection of benchmark datasets for learning with graphs.
Christopher Morris,Nils M. Kriege,Franka Bause,Kristian Kersting,Petra Mutzel,Marion Neumann +5 more
TL;DR: The TUDataset for graph classification and regression is introduced, which consists of over 120 datasets of varying sizes from a wide range of applications and provides Python-based data loaders, kernel and graph neural network baseline implementations, and evaluation tools.
Journal ArticleDOI
Building attention and edge message passing neural networks for bioactivity and physical–chemical property prediction
TL;DR: This work removes the need to introduce a priori knowledge of the task and chemical descriptor calculation by using only fundamental graph-derived properties, and sets a new standard on sparse multi-task virtual screening targets.
Proceedings ArticleDOI
SUGAR: Subgraph Neural Network with Reinforcement Pooling and Self-Supervised Mutual Information Mechanism
TL;DR: Wang et al. as discussed by the authors proposed a hierarchical subgraph-level selection and embedding-based graph neural network for graph classification, which can learn more discriminative subgraph representations and respond in an explanatory way.
Posted ContentDOI
Graph Kernels: A Survey.
TL;DR: This survey presents a comprehensive overview of a wide range of graph kernels, and performs an experimental evaluation of several of those kernels on publicly available datasets, and provides a comparative study.
References
More filters
Journal ArticleDOI
LIBSVM: A library for support vector machines
Chih-Chung Chang,Chih-Jen Lin +1 more
TL;DR: Issues such as solving SVM optimization problems theoretical convergence multiclass classification probability estimates and parameter selection are discussed in detail.
Journal ArticleDOI
Support-Vector Networks
Corinna Cortes,Vladimir Vapnik +1 more
TL;DR: High generalization ability of support-vector networks utilizing polynomial input transformations is demonstrated and the performance of the support- vector network is compared to various classical learning algorithms that all took part in a benchmark study of Optical Character Recognition.
Journal Article
Visualizing Data using t-SNE
TL;DR: A new technique called t-SNE that visualizes high-dimensional data by giving each datapoint a location in a two or three-dimensional map, a variation of Stochastic Neighbor Embedding that is much easier to optimize, and produces significantly better visualizations by reducing the tendency to crowd points together in the center of the map.
Posted Content
Efficient Estimation of Word Representations in Vector Space
TL;DR: This paper proposed two novel model architectures for computing continuous vector representations of words from very large data sets, and the quality of these representations is measured in a word similarity task and the results are compared to the previously best performing techniques based on different types of neural networks.
Posted Content
Semi-Supervised Classification with Graph Convolutional Networks
Thomas Kipf,Max Welling +1 more
TL;DR: A scalable approach for semi-supervised learning on graph-structured data that is based on an efficient variant of convolutional neural networks which operate directly on graphs which outperforms related methods by a significant margin.