scispace - formally typeset
Open AccessJournal ArticleDOI

A survey on graph kernels

Reads0
Chats0
TLDR
This survey gives a comprehensive overview of techniques for kernel-based graph classification developed in the past 15 years and describes and categorizes graph kernels based on properties inherent to their design, such as the nature of their extracted graph features, their method of computation and their applicability to problems in practice.
Abstract
Graph kernels have become an established and widely-used technique for solving classification tasks on graphs. This survey gives a comprehensive overview of techniques for kernel-based graph classification developed in the past 15 years. We describe and categorize graph kernels based on properties inherent to their design, such as the nature of their extracted graph features, their method of computation and their applicability to problems in practice. In an extensive experimental evaluation, we study the classification accuracy of a large suite of graph kernels on established benchmarks as well as new datasets. We compare the performance of popular kernels with several baseline methods and study the effect of applying a Gaussian RBF kernel to the metric induced by a graph kernel. In doing so, we find that simple baselines become competitive after this transformation on some datasets. Moreover, we study the extent to which existing graph kernels agree in their predictions (and prediction errors) and obtain a data-driven categorization of kernels as result. Finally, based on our experimental results, we derive a practitioner’s guide to kernel-based graph classification.

read more

Content maybe subject to copyright    Report

Citations
More filters
Book

Graph Representation Learning

TL;DR: This work has shown that graph-structured data is ubiquitous throughout the natural and social sciences, from telecommunication networks to quantum chemistry, and building relational inductive biases into deep learnin...
Posted Content

TUDataset: A collection of benchmark datasets for learning with graphs.

TL;DR: The TUDataset for graph classification and regression is introduced, which consists of over 120 datasets of varying sizes from a wide range of applications and provides Python-based data loaders, kernel and graph neural network baseline implementations, and evaluation tools.
Journal ArticleDOI

Building attention and edge message passing neural networks for bioactivity and physical–chemical property prediction

TL;DR: This work removes the need to introduce a priori knowledge of the task and chemical descriptor calculation by using only fundamental graph-derived properties, and sets a new standard on sparse multi-task virtual screening targets.
Proceedings ArticleDOI

SUGAR: Subgraph Neural Network with Reinforcement Pooling and Self-Supervised Mutual Information Mechanism

TL;DR: Wang et al. as discussed by the authors proposed a hierarchical subgraph-level selection and embedding-based graph neural network for graph classification, which can learn more discriminative subgraph representations and respond in an explanatory way.
Posted ContentDOI

Graph Kernels: A Survey.

TL;DR: This survey presents a comprehensive overview of a wide range of graph kernels, and performs an experimental evaluation of several of those kernels on publicly available datasets, and provides a comparative study.
References
More filters
Journal ArticleDOI

LIBSVM: A library for support vector machines

TL;DR: Issues such as solving SVM optimization problems theoretical convergence multiclass classification probability estimates and parameter selection are discussed in detail.
Journal ArticleDOI

Support-Vector Networks

TL;DR: High generalization ability of support-vector networks utilizing polynomial input transformations is demonstrated and the performance of the support- vector network is compared to various classical learning algorithms that all took part in a benchmark study of Optical Character Recognition.
Journal Article

Visualizing Data using t-SNE

TL;DR: A new technique called t-SNE that visualizes high-dimensional data by giving each datapoint a location in a two or three-dimensional map, a variation of Stochastic Neighbor Embedding that is much easier to optimize, and produces significantly better visualizations by reducing the tendency to crowd points together in the center of the map.
Posted Content

Efficient Estimation of Word Representations in Vector Space

TL;DR: This paper proposed two novel model architectures for computing continuous vector representations of words from very large data sets, and the quality of these representations is measured in a word similarity task and the results are compared to the previously best performing techniques based on different types of neural networks.
Posted Content

Semi-Supervised Classification with Graph Convolutional Networks

TL;DR: A scalable approach for semi-supervised learning on graph-structured data that is based on an efficient variant of convolutional neural networks which operate directly on graphs which outperforms related methods by a significant margin.
Related Papers (5)