scispace - formally typeset
Search or ask a question
Author

Philipp Trunschke

Bio: Philipp Trunschke is an academic researcher from Technical University of Berlin. The author has contributed to research in topics: Mathematics & Partial differential equation. The author has an hindex of 4, co-authored 10 publications receiving 48 citations.

Papers
More filters
Journal ArticleDOI
TL;DR: In this paper, a statistical learning approach for high-dimensional parametric PDEs related to uncertainty quantification is derived, based on the minimization of an empirical risk on a selected model class, and it is shown to be applicable to a broad range of problems.
Abstract: A statistical learning approach for high-dimensional parametric PDEs related to uncertainty quantification is derived. The method is based on the minimization of an empirical risk on a selected model class, and it is shown to be applicable to a broad range of problems. A general unified convergence analysis is derived, which takes into account the approximation and the statistical errors. By this, a combination of theoretical results from numerical analysis and statistics is obtained. Numerical experiments illustrate the performance of the method with the model class of hierarchical tensors.

28 citations

Journal ArticleDOI
TL;DR: A basic approach in defining best practice procedures of measuring consistent data sets in heterogeneous catalysis using “handbooks” was presented and Selective oxidation of short-chain alkanes over mixed metal oxide catalysts was selected as an example.
Abstract: The “Seven Pillars” of oxidation catalysis proposed by Robert K. Grasselli represent an early example of phenomenological descriptors in the field of heterogeneous catalysis. Major advances in the theoretical description of catalytic reactions have been achieved in recent years and new catalysts are predicted today by using computational methods. To tackle the immense complexity of high-performance systems in reactions where selectivity is a major issue, analysis of scientific data by artificial intelligence and data science provides new opportunities for achieving improved understanding. Modern data analytics require data of highest quality and sufficient diversity. Existing data, however, frequently do not comply with these constraints. Therefore, new concepts of data generation and management are needed. Herein we present a basic approach in defining best practice procedures of measuring consistent data sets in heterogeneous catalysis using “handbooks”. Selective oxidation of short-chain alkanes over mixed metal oxide catalysts was selected as an example.

26 citations

Posted Content
TL;DR: The results are based on a restricted isometry property (RIP) which holds in probability and is independent of the nonlinear least squares setting and demonstrate the advantage of an optimal sampling density for sets of functions with sparse representations.
Abstract: We consider best approximation problems in a nonlinear subset $\mathcal{M}$ of a Banach space of functions $(\mathcal{V},\|\bullet\|)$. The norm is assumed to be a generalization of the $L^2$-norm for which only a weighted Monte Carlo estimate $\|\bullet\|_n$ can be computed. The objective is to obtain an approximation $v\in\mathcal{M}$ of an unknown function $u \in \mathcal{V}$ by minimizing the empirical norm $\|u-v\|_n$. We consider this problem for general nonlinear subsets and establish error bounds for the empirical best approximation error. Our results are based on a restricted isometry property (RIP) which holds in probability and is independent of the nonlinear least squares setting. Several model classes are examined where analytical statements can be made about the RIP and the results are compared to existing sample complexity bounds from the literature. We find that for well-studied model classes our general bound is weaker but exhibits many of the same properties as these specialized bounds. Notably, we demonstrate the advantage of an optimal sampling density (as known for linear spaces) for sets of functions with sparse representations.

12 citations

Posted Content
TL;DR: In this paper, an efficient compression technique based on hierarchical tensors for popular option pricing methods is presented, which can be used for the computation of Bermudan option prices with the Monte Carlo least-squares approach as well as the dual martingale method, both using high-dimensional tensorized polynomial expansions.
Abstract: An efficient compression technique based on hierarchical tensors for popular option pricing methods is presented. It is shown that the "curse of dimensionality" can be alleviated for the computation of Bermudan option prices with the Monte Carlo least-squares approach as well as the dual martingale method, both using high-dimensional tensorized polynomial expansions. This discretization allows for a simple and computationally cheap evaluation of conditional expectations. Complexity estimates are provided as well as a description of the optimization procedures in the tensor train format. Numerical experiments illustrate the favourable accuracy of the proposed methods. The dynamical programming method yields results comparable to recent Neural Network based methods.

5 citations

Journal ArticleDOI
TL;DR: This work proposes to extend the low-rank tensors framework by including the concept of block-sparsity, in the context of polynomial regression, to adapt the ansatz space to align better with known sample complexity results.
Abstract: Low-rank tensor are an established framework for the parametrization of multivariate polynomials. We propose to extend this framework by including the concept of block-sparsity to efficiently parametrize homogeneous, multivariate polynomials with low-rank tensors. This provides a representation of general multivariate polynomials as a sum of homogeneous, multivariate polynomials, represented by block-sparse, low-rank tensors. We show that this sum can be concisely represented by a single block-sparse, low-rank tensor. We further prove cases, where low-rank tensors are particularly well suited by showing that for banded symmetric tensors of homogeneous polynomials the block sizes in the block-sparse multivariate polynomial space can be bounded independent of the number of variables. We showcase this format by applying it to high-dimensional least squares regression problems where it demonstrates improved computational resource utilization and sample efficiency.

4 citations


Cited by
More filters
Journal ArticleDOI
TL;DR: In this paper, the authors provide a survey of recent scientific works that incorporate machine learning and the way that explainable machine learning is used in combination with domain knowledge from the application areas.
Abstract: Machine learning methods have been remarkably successful for a wide range of application areas in the extraction of essential information from data. An exciting and relatively recent development is the uptake of machine learning in the natural sciences, where the major goal is to obtain novel scientific insights and discoveries from observational or simulated data. A prerequisite for obtaining a scientific outcome is domain knowledge, which is needed to gain explainability, but also to enhance scientific consistency. In this article, we review explainable machine learning in view of applications in the natural sciences and discuss three core elements that we identified as relevant in this context: transparency, interpretability, and explainability. With respect to these core elements, we provide a survey of recent scientific works that incorporate machine learning and the way that explainable machine learning is used in combination with domain knowledge from the application areas.

493 citations

Journal Article
TL;DR: A trustful prediction of new promising materials, identification of anomalies, and scientific advancement are doubtful when the scientific connection between the descriptor and the actuating mechanisms is unclear.
Abstract: Statistical learning of materials properties or functions so far starts with a largely silent, nonchallenged step: the choice of the set of descriptive parameters (termed descriptor). However, when the scientific connection between the descriptor and the actuating mechanisms is unclear, the causality of the learned descriptor-property relation is uncertain. Thus, a trustful prediction of new promising materials, identification of anomalies, and scientific advancement are doubtful. We analyze this issue and define requirements for a suitable descriptor. For a classic example, the energy difference of zinc blende or wurtzite and rocksalt semiconductors, we demonstrate how a meaningful descriptor can be found systematically.

455 citations

Journal ArticleDOI
TL;DR: The existence of a small reduced basis is used to construct neural networks that yield approximations of the parametric solution maps in such a way that the sizes of these networks essentially only depend on the size of the reduced basis.
Abstract: We derive upper bounds on the complexity of ReLU neural networks approximating the solution maps of parametric partial differential equations. In particular, without any knowledge of its concrete shape, we use the inherent low dimensionality of the solution manifold to obtain approximation rates which are significantly superior to those provided by classical neural network approximation results. Concretely, we use the existence of a small reduced basis to construct, for a large variety of parametric partial differential equations, neural networks that yield approximations of the parametric solution maps in such a way that the sizes of these networks essentially only depend on the size of the reduced basis.

133 citations

Posted Content
TL;DR: In this article, the authors derive upper bounds on the complexity of ReLU neural networks approximating the solution maps of parametric partial differential equations, without any knowledge of its concrete shape, using the inherent low-dimensionality of the solution manifold to obtain approximation rates which are significantly superior to those provided by classical neural networks.
Abstract: We derive upper bounds on the complexity of ReLU neural networks approximating the solution maps of parametric partial differential equations. In particular, without any knowledge of its concrete shape, we use the inherent low-dimensionality of the solution manifold to obtain approximation rates which are significantly superior to those provided by classical neural network approximation results. Concretely, we use the existence of a small reduced basis to construct, for a large variety of parametric partial differential equations, neural networks that yield approximations of the parametric solution maps in such a way that the sizes of these networks essentially only depend on the size of the reduced basis.

77 citations

Posted Content
TL;DR: This work finds strong support for the hypothesis that approximation-theoretical effects heavily influence the practical behavior of learning problems in numerical analysis.
Abstract: We perform a comprehensive numerical study of the effect of approximation-theoretical results for neural networks on practical learning problems in the context of numerical analysis. As the underlying model, we study the machine-learning-based solution of parametric partial differential equations. Here, approximation theory predicts that the performance of the model should depend only very mildly on the dimension of the parameter space and is determined by the intrinsic dimension of the solution manifold of the parametric partial differential equation. We use various methods to establish comparability between test-cases by minimizing the effect of the choice of test-cases on the optimization and sampling aspects of the learning problem. We find strong support for the hypothesis that approximation-theoretical effects heavily influence the practical behavior of learning problems in numerical analysis.

48 citations