scispace - formally typeset
Search or ask a question

Showing papers presented at "The European Symposium on Artificial Neural Networks in 2019"


Proceedings Article
01 Jan 2019
TL;DR: A recurrent Deep Learning based model to generate graphs by learning to predict their ordered edge sequence is proposed, outperforming canonical graph generative models from graph theory, and reaching performances comparable to the current state of the art on graph generation.
Abstract: Graph generation with Machine Learning models is a challenging problem with applications in various research fields. Here, we propose a recurrent Deep Learning based model to generate graphs by learning to predict their ordered edge sequence. Despite its simplicity, our experiments on a wide range of datasets show that our approach is able to generate graphs originating from very different distributions, outperforming canonical graph generative models from graph theory, and reaching performances comparable to the current state of the art on graph generation.

17 citations


Proceedings Article
24 Apr 2019
TL;DR: In this article, the feasibility of using synthetic medical data generated by GANs in the classroom, to teach data science in health infor-matics is examined, and an end-to-end methodology to retain instructional utility, while preserving privacy to a level, which meets regulatory requirements is presented.
Abstract: We examine the feasibility of using synthetic medical data generated by GANs in the classroom, to teach data science in health infor-matics. We present an end-to-end methodology to retain instructional utility, while preserving privacy to a level, which meets regulatory requirements: (1) a GAN is trained by a certified medical-data security-aware agent, inside a secure environment; (2) the final GAN model is used outside of the secure environment by external users (instructors or researchers) to generate synthetic data. This second step facilitates data handling for external users, by avoiding de-identification, which may require special user training, be costly, and/or cause loss of data fidelity. We benchmark our proposed GAN versus various baseline methods using a novel set of metrics. At equal levels of privacy and utility, GANs provide small footprint models, meeting the desired specifications of our application domain. Data, code, and a challenge that we organized for educational purposes are available.

17 citations


Proceedings Article
01 Jan 2019
TL;DR: This work unify the policy and the feedback to favor actions of low probability density and shows that the proposed approach increases the accumulated reward in comparison to the autonomous learning method.
Abstract: Reinforcement learning is utilized in contexts where an agent tries to learn from the environment. Using continuous actions, the performance may be improved in comparison to using discrete actions, however, this leads to excessive time to find a proper policy. In this work, we focus on including human feedback in reinforcement learning for a continuous action space. We unify the policy and the feedback to favor actions of low probability density. Furthermore, we compare the performance of the feedback for the continuous actor-critic algorithm and test our experiments in the cart-pole balancing task. The obtained results show that the proposed approach increases the accumulated reward in comparison to the autonomous learning method.

16 citations


Proceedings Article
01 Jan 2019
TL;DR: The Deep Embedded Self-Organizing Map is introduced, a model that jointly learns representations and the code vectors of a self-organizing map that is composed of an autoencoder and a custom SOM layer that are optimized in a joint training procedure.
Abstract: In the wake of recent advances in joint clustering and deep learning, we introduce the Deep Embedded Self-Organizing Map, a model that jointly learns representations and the code vectors of a self-organizing map. Our model is composed of an autoencoder and a custom SOM layer that are optimized in a joint training procedure, motivated by the idea that the SOM prior could help learning SOM-friendly representations. We evaluate SOM-based models in terms of clustering quality and unsupervised clustering accuracy, and study the bene ts of joint training.

14 citations


Proceedings Article
01 Jan 2019
TL;DR: In this paper, a detection method based on combining nonlinear dimensionality reduction and density estimation techniques was proposed to detect adversarial examples crafted by non-adaptive attackers, i.e., not specifically tuned to bypass the detection method.
Abstract: Deep neural networks are vulnerable to adversarial examples, i.e., carefully-perturbed inputs aimed to mislead classification. This work proposes a detection method based on combining non-linear dimensionality reduction and density estimation techniques. Our empirical findings show that the proposed approach is able to effectively detect adversarial examples crafted by non-adaptive attackers, i.e., not specifically tuned to bypass the detection method. Given our promising results, we plan to extend our analysis to adaptive attackers in future work.

13 citations


Proceedings Article
24 Jan 2019
TL;DR: Several interpolation schemes on the manifold of fixed-rank positive-semidefinite (PSD) matrices are presented and how these techniques can be used for model order reduction of parameterized linear dynamical systems are explained.
Abstract: We present several interpolation schemes on the manifold of fixed-rank positive-semidefinite (PSD) matrices. We explain how these techniques can be used for model order reduction of parameterized linear dynamical systems, and obtain preliminary results on an application.

13 citations


Proceedings Article
01 Jan 2019
TL;DR: This work proposes to exploit the representational power of sparse dictionaries to determine image local properties that can be used as crucial ingredients of humanly understandable explanations of classification decisions.
Abstract: A pressing research topic is to find ways to explain the decisions of machine learning systems to end users, data officers, and other stakeholders. These explanations must be understandable to human beings. Much work in this field focuses on image classification, as the required explanations can rely on images, therefore making communication relatively easy, and may take into account the image as a whole. Here, we propose to exploit the representational power of sparse dictionaries to determine image local properties that can be used as crucial ingredients of humanly understandable explanations of classification decisions.

11 citations


Proceedings Article
01 Jan 2019
TL;DR: This tutorial provides an introduction into the field of streaming data analysis summarizing its major characteristics and highlighting important research directions in the analysis of dynamic data.
Abstract: Today, many data are not any longer static but occur as dynamic data streams with high velocity, variability and volume. This leads to new challenges to be addressed by novel or adapted algorithms. In this tutorial we provide an introduction into the field of streaming data analysis summarizing its major characteristics and highlighting important research directions in the analysis of dynamic data.

10 citations


Proceedings Article
24 Apr 2019
TL;DR: This work considers the diagonalization criterion studied in a seminal paper by Pham (2001), and proposes a new quasi-Newton method for its optimization that outperforms Pham's algorithm.
Abstract: The approximate joint diagonalization of a set of matrices consists in finding a basis in which these matrices are as diagonal as possible. This problem naturally appears in several statistical learning tasks such as blind signal separation. We consider the diagonalization criterion studied in a seminal paper by Pham (2001), and propose a new quasi-Newton method for its optimization. Through numerical experiments on simulated and real datasets, we show that the proposed method outper-forms Pham's algorithm. An open source Python package is released.

9 citations


Proceedings Article
01 Jan 2019
TL;DR: Frequency Domain Transformer Network (FDTN) as mentioned in this paper is an end-to-end learnable model that estimates and uses the transformations of the signal in the frequency domain to predict the next frames given some previous frames.
Abstract: The task of video prediction is forecasting the next frames given some previous frames. Despite much recent progress, this task is still challenging mainly due to high nonlinearity in the spatial domain. To address this issue, we propose a novel architecture, Frequency Domain Transformer Network (FDTN), which is an end-to-end learnable model that estimates and uses the transformations of the signal in the frequency domain. Experimental evaluations show that this approach can outperform some widely used video prediction methods like Video Ladder Network (VLN) and Predictive Gated Pyramids (PGP).

9 citations


Proceedings Article
01 Jan 2019
TL;DR: A concept drift detection strategy followed by a prototype based insertion strategy is proposed that provides stable and quick adjustments in times of change.
Abstract: Todays datasets, especially in streaming context, are more and more non-static and require algorithms to detect and adapt to change. Recent work shows vital research in the field, but mainly lack stable performance during model adaptation. In this work, a concept drift detection strategy followed by a prototype based insertion strategy is proposed. Validated through experimental results on a variety of typical non-static data, our solution provides stable and quick adjustments in times of change.

Proceedings Article
01 Jul 2019
TL;DR: This article proposed a gradient-based optimization strategy to generate a symmetric mixture of Gaussian modes (SGM) where each mode belongs to a particular quantization stage and achieved 2-bit state-of-the-art performance.
Abstract: Due to their high computational complexity, deep neural networks are still limited to powerful processing units. To promote a reduced model complexity by dint of low-bit fixed-point quantization, we propose a gradient-based optimization strategy to generate a symmetric mixture of Gaussian modes (SGM) where each mode belongs to a particular quantization stage. We achieve 2-bit state-of-the-art performance and illustrate the model's ability for self-dependent weight adaptation during training.

Proceedings Article
01 Jan 2019
TL;DR: This paper proposed a variational auto-encoder prior with a weakly informative multivariate Student's t-distribution, which allows for a more robust approximation of the underlying data distribution.
Abstract: We propose a new structure for the variational auto-encoders (VAEs) prior, with the weakly informative multivariate Student’s t-distribution. In the proposed model all distribution parameters are trained, thereby allowing for a more robust approximation of the underlying data distribution. We used Fashion-MNIST data in two experiments to compare the proposed VAEs with the standard Gaussian priors. Both experiments showed a better reconstruction of the images with VAEs using Student’s t-prior distribution.

Proceedings Article
24 Apr 2019
TL;DR: In this paper, a neural network embedding approach is proposed to model power transmission grids, in which high voltage lines are disconnected and reconnected with one-another from time to time, either accidentally or willfully.
Abstract: We propose a novel neural network embedding approach to model power transmission grids, in which high voltage lines are disconnected and reconnected with one-another from time to time, either accidentally or willfully. We call our architeture LEAP net, for Latent Encoding of Atypical Perturbation. Our method implements a form of transfer learning, permitting to train on a few source domains, then generalize to new target domains, without learning on any example of that domain. We evaluate the viability of this technique to rapidly assess cu-rative actions that human operators take in emergency situations, using real historical data, from the French high voltage power grid.

Proceedings Article
01 Jan 2019
TL;DR: Through empirical evalution on diverse benchmark datasets from the UCR Time Series Classification (TSC) Archive, it is shown that classifiers trained on SPGF-TimeNet-based hybrid and generic features outperform state-of-the-art TSC algorithms such as BOSS, while being computationally efficient.
Abstract: Automated feature extraction from time series to capture statistical, temporal, spectral, and morphololgical properties is highly desirable but challenging due to diverse nature of real-world time series applications. In this paper, we consider extracting a rich and robust set of time series features encompassing signal processing based features as well as generic hierarchical features extracted via deep neural networks. We present SPGF-TimeNet : a generic feature extractor for time series that allows fusion of signal processing, information-theoretic, and statistical features (Signal Properties based Generic Features (SPGF )) with features from an off-the-shelf pre-trained deep recurrent neural network (TimeNet). Through empirical evalution on diverse benchmark datasets from the UCR Time Series Classification (TSC) Archive, we show that classifiers trained on SPGF-TimeNet-based hybrid and generic features outperform state-of-the-art TSC algorithms such as BOSS, while being computationally efficient.

Proceedings Article
01 Jan 2019
TL;DR: A high-level overview of the state-of-the-art in representation learning and embeddings for structured data across a wide range of machine learning fields can be found in this paper.
Abstract: Performing machine learning on structured data is complicated by the fact that such data does not have vectorial form. Therefore, multiple approaches have emerged to construct vectorial representations of structured data, from kernel and distance approaches to recurrent, recursive, and convolutional neural networks. Recent years have seen heightened attention in this demanding field of research and several new approaches have emerged, such as metric learning on structured data, graph convolutional neural networks, and recurrent decoder networks for structured data. In this contribution, we provide an high-level overview of the state-of-the-art in representation learning and embeddings for structured data across a wide range of machine learning fields.

Proceedings Article
01 Jan 2019
TL;DR: Experimental results show that the proposed class-aware t-S NE (cat-SNE) outperforms regular t- SNE in KNN classification tasks carried out in the embedding.
Abstract: Stochastic Neighbor Embedding (SNE) and variants like t-distributed SNE are popular methods of unsupervised dimensionality reduction (DR) that deliver outstanding experimental results. Regular tSNE is often used to visualize data with class labels in colored scatterplots, even if those labels are actually not involved in the DR process. This paper proposes a modification of t-SNE that employs class labels to adjust the widths of the Gaussian neighborhoods around each datum, instead of deriving those from a perplexity set by the user. The widths are fixed to concentrate a major fraction of the probability distribution around a datum on neighbors with the same class. This tends to shrink the bulk of the classes and to stretch their low-dimensional separation. Experimental results show that the proposed class-aware t-SNE (cat-SNE) outperforms regular t-SNE in KNN classification tasks carried out in the embedding.

Proceedings Article
01 Jan 2019
TL;DR: The design of a convolutional core that utilizes an approximate log multiplier to significantly reduce the power consumption of FPGA acceleration of Convolutional neural networks and exploits FPGAs reconfigurability as well as the parallelism and input sharing opportunities in convolutionAL layers to minimize the costs.
Abstract: This paper presents the design of a convolutional core that utilizes an approximate log multiplier to significantly reduce the power consumption of FPGA acceleration of convolutional neural networks. The core also exploits FPGA reconfigurability as well as the parallelism and input sharing opportunities in convolutional layers to minimize the costs. The simulation results show reductions up to 78.19% of LUT usage and 60.54% of power consumption compared to the core that uses exact fixed-point multiplier, while maintaining comparable accuracy on a subset of MNIST dataset.

Proceedings Article
01 Jan 2019
TL;DR: Regression WiSARD and ClusRegressionWiSARD, two new weightless neural network models that were applied in the challenging task of predicting the total palm oil production of a set of 28 differently located sites under different climate and soil profiles are introduced.
Abstract: This paper introduces Regression WiSARD and ClusRegression WiSARD, two new weightless neural network models that were applied in the challenging task of predicting the total palm oil production of a set of 28 differently located sites under different climate and soil profiles. Both models were derived from the n-tuple regression weightless neural model and obtained error (MAE) rates of 0.08737% and 0.08938%, respectively, which are very competitive with the state-of-art (0.07569), whilst being four (4) orders of magnitude faster during the training phase.

Proceedings Article
01 Apr 2019
TL;DR: Statistical physics methods complement other approaches to the theoretical understanding of machine learning processes and inference in stochastic modeling and inspire novel and efficient algorithms and facilitates interdisciplinary applications in a variety of scientific and technical disciplines.
Abstract: The exchange of ideas between statistical physics and computer science has been very fruitful and is currently gaining momentum as a consequence of the revived interest in neural networks, machine learning and inference in general Statistical physics methods complement other approaches to the theoretical understanding of machine learning processes and inference in stochastic modeling They facilitate, for instance, the study of dynamical and equilibrium properties of randomized training processes in model situations At the same time, the approach inspires novel and efficient algorithms and facilitates interdisciplinary applications in a variety of scientific and technical disciplines

Proceedings Article
01 Jun 2019
TL;DR: This work develops automatic techniques with special focus on detecting such ghostwriting in high school assignments by training deep neural networks on an unprecedented large amount of data supplied by the Danish company MaCom, which covers 90% of Danish high schools.
Abstract: Students hiring ghostwriters to write their assignments is an increasing problem in educational institutions all over the world, with companies selling these services as a product. In this work, we develop automatic techniques with special focus on detecting such ghostwriting in high school assignments. This is done by training deep neural networks on an unprecedented large amount of data supplied by the Danish company MaCom, which covers 90% of Danish high schools. We achieve an accuracy of 0.875 and a AUC score of 0.947 on an evenly split data set.

Proceedings Article
01 Jan 2019
TL;DR: This work develops several neural networks (NN) for classifying images in Fourier domain and visualize neuron activities and analyze the underlying patterns relevant for the learning and classification process.
Abstract: Image classification is successfully done with Convolutional Neural Networks (CNN). Alternatively it can be done in Fourier domain avoiding the convolution process. In this work, we develop several neural networks (NN) for classifying images in Fourier domain. In order to understand and explain the behaviour of the built NNs, we visualize neuron activities and analyze the underlying patterns relevant for the learning and classification process. We have carried out comparative study based on several datasets. By using images of objects with partial occlusion, we are able to find out the parts that are important for the classification of certain objects.

Proceedings Article
01 Jan 2019
TL;DR: This work proposes a new approach to robotic grasping exploiting conditional Wasserstein generative adversarial networks (WGANs), which output promising grasp candidates from depth image inputs, and compares it to a classical grasp planner for primitive geometric object shapes and a state-of-the-art discriminative network model.
Abstract: This work proposes a new approach to robotic grasping exploiting conditional Wasserstein generative adversarial networks (WGANs), which output promising grasp candidates from depth image inputs. In contrast to discriminative models, the WGAN approach enables deliberative navigation in the set of feasible grasps and thus allows a smooth integration with other motion planning tools. We find that the training autonomously partitioned the space of feasible grasps into several regions corresponding to different grasp types. Each region forms a smooth grasp manifold with latent parameters corresponding to important grasp parameters like approach direction. We evaluate the model in simulation on the multi-fingered Shadow Robot hand, comparing it a) to a classical grasp planner for primitive geometric object shapes and b) to a state-of-the-art discriminative network model. The proposed generative model matches the grasp success rate of its trainer models and exhibits better generalization.

Proceedings Article
28 Mar 2019
TL;DR: An optimal attack strategy against online learning classifiers is formulated and two defence mechanisms to mitigate the effect of online poisoning attacks are proposed by analysing the impact of the data points in the classifier and by means of an adaptive combination of machine learning classifier with different learning rates.
Abstract: Machine learning systems are vulnerable to data poisoning, a coordinated attack where a fraction of the training dataset is manipulated by an attacker to subvert learning. In this paper we first formulate an optimal attack strategy against online learning classifiers to assess worst-case scenarios. We also propose two defence mechanisms to mitigate the effect of online poisoning attacks by analysing the impact of the data points in the classifier and by means of an adaptive combination of machine learning classifiers with different learning rates. Our experimental evaluation supports the usefulness of our proposed defences to mitigate the effect of poisoning attacks in online learning settings.

Proceedings Article
01 Jan 2019
TL;DR: This work considers the problem of active one-shot classification where a classifier needs to adapt to new tasks by requesting labels for one example per class from (potentially many) unlabeled examples and proposes a clustering approach to the problem.
Abstract: We consider the problem of active one-shot classification where a classifier needs to adapt to new tasks by requesting labels for one example per class from (potentially many) unlabeled examples. We propose a clustering approach to the problem. The features extracted with Prototypical Networks [1] are clustered using K-means and the label for one representative sample from each cluster is requested to label the whole cluster. We demonstrate good performance of this simple active adaptation strategy using image data.

Proceedings Article
01 Jan 2019
TL;DR: It is shown that, even where technological solutions are available, the law needs to keep up to support and accurately regulate the use of the technological solutions and to identify stumble points in this regard.
Abstract: This paper discusses whether the law is up to regulate machine learning (”ML”) model-based decision-making in the context of the railways. We especially deal with the fairness and accountability of these models when exploited in the context of train traffic management (”TTM”). Railway sector-specific regulation, in their quality as network industry, hereby serves as a pilot. We show that, even where technological solutions are available, the law needs to keep up to support and accurately regulate the use of the technological solutions and we identify stumble points in this regard.

Proceedings Article
20 Feb 2019
TL;DR: This work aims for an identification of feature relevance bounds which - besides identifying all relevant features - explicitly differentiates between strongly and weakly relevant features.
Abstract: The increasing occurrence of ordinal data, mainly sociodemographic, led to a renewed research interest in ordinal regression, i.e. the prediction of ordered classes. Besides model accuracy, the interpretation of these models itself is of high relevance, and existing approaches therefore enforce e.g. model sparsity. For high dimensional or highly correlated data, however, this might be misleading due to strong variable dependencies. In this contribution, we aim for an identification of feature relevance bounds which - besides identifying all relevant features - explicitly differentiates between strongly and weakly relevant features.

Proceedings Article
01 Mar 2019
TL;DR: The state-of-the-art method RefineNet is adapted for 3D segmentation of the plant root MRI images in super-resolution, which contains most root structures, including branches not extracted by the human annotator.
Abstract: Analyzing plant roots is crucial to understand plant performance in different soil environments. While magnetic resonance imaging (MRI) can be used to obtain 3D images of plant roots, extracting the root structural model is challenging due to highly noisy soil environments and low-resolution of MRI images. To improve both contrast and resolution, we adapt the state-of-the-art method RefineNet for 3D segmentation of the plant root MRI images in super-resolution. The networks are trained from few manual segmentations that are augmented by geometric transformations, realistic noise, and other variabilities. The resulting segmentations contain most root structures, including branches not extracted by the human annotator.

Proceedings Article
01 Jan 2019
TL;DR: A new network architecture, the Blind-Spot Network, is proposed that, according to the presented experiments, improves the performance of autoencoders in this setting.
Abstract: The development of computer-aided screening (CAS) systems is motivated by the high prevalence and severity of the target disease along with the time taken to manually assess each case. This is the case with diabetic retinopathy screening, that is based on the manual grading of retinography images. The development of CAS systems, however, usually involves data-driven approaches that require extensive and usually scarce manually labeled datasets. With this in mind, we propose the use of unsupervised anomaly detection methods for screening that can take advantage of the large amount of healthy cases available. Concretely, we focus on reconstruction-based anomaly detection methods, which are usually approached with autoencoders. We propose a new network architecture, the Blind-Spot Network, that, according to the presented experiments, improves the performance of autoencoders in this setting.

Proceedings Article
01 Jan 2019
TL;DR: A strategy of using Bluetooth Low Energy beacons for device-free presence detection is introduced and a distinction is made between so-called devicebased active and device- free passive sensing systems.
Abstract: In an era of smart information systems and smart buildings, detecting, tracking and identifying the presence of attendants inside of enclosed rooms have evolved to a key challenge in the research area of smart building systems. Therefore, several types of sensing systems were proposed over the past decade to tackle these challenge. Depending on the component’s arrangement, a distinction is made between so-called devicebased active and device-free passive sensing systems. Here we focus on the device-free passive concept and introduce a strategy of using Bluetooth Low Energy beacons for device-free presence detection.