scispace - formally typeset
Search or ask a question
Journal ArticleDOI

Convolutional Neural Networks for Long Time Dissipative Quantum Dynamics.

05 Mar 2021-Journal of Physical Chemistry Letters (American Chemical Society (ACS))-Vol. 12, Iss: 9, pp 2476-2483
TL;DR: In this article, a deep artificial neural network composed of convolutional layers is used for predicting long-time dynamics of open quantum systems provided the preceding short-time evolution of a system is known.
Abstract: Exact numerical simulations of dynamics of open quantum systems often require immense computational resources. We demonstrate that a deep artificial neural network composed of convolutional layers is a powerful tool for predicting long-time dynamics of open quantum systems provided the preceding short-time evolution of a system is known. The neural network model developed in this work simulates long-time dynamics efficiently and accurately across different dynamical regimes from weakly damped coherent motion to incoherent relaxation. The model was trained on a data set relevant to photosynthetic excitation energy transfer and can be deployed to study long-lasting quantum coherence phenomena observed in light-harvesting complexes. Furthermore, our model performs well for the initial conditions different than those used in the training. Our approach reduces the required computational resources for long-time simulations and holds the promise for becoming a valuable tool in the study of open quantum systems.
Citations
More filters
Journal ArticleDOI
20 May 2021
TL;DR: A broad range of machine learning applications in excited-state research can be found in this article, which includes the prediction of molecular properties, improvements of quantum mechanical methods for the calculations of excited state properties and the search for new materials.
Abstract: Theoretical simulations of electronic excitations and associated processes in molecules are indispensable for fundamental research and technological innovations. However, such simulations are notoriously challenging to perform with quantum mechanical methods. Advances in machine learning open many new avenues for assisting molecular excited-state simulations. In this Review, we track such progress, assess the current state of the art and highlight the critical issues to solve in the future. We overview a broad range of machine learning applications in excited-state research, which include the prediction of molecular properties, improvements of quantum mechanical methods for the calculations of excited-state properties and the search for new materials. Machine learning approaches can help us understand hidden factors that influence photo-processes, leading to a better control of such processes and new rules for the design of materials for optoelectronic applications. Machine learning is starting to reshape our approaches to excited-state simulations by accelerating and improving or even completely bypassing traditional theoretical methods. It holds big promises for taking the optoelectronic materials design to a new level.

73 citations

Journal ArticleDOI
TL;DR: In this article, a large number of LSTM-NNs are constructed by resampling time-series sequences that were obtained from the early stage quantum evolution given by numerically exact multilayer multiconfigurational time dependent Hartree method.
Abstract: The recurrent neural network with the long short-term memory cell (LSTM-NN) is employed to simulate the long-time dynamics of open quantum systems. The bootstrap method is applied in the LSTM-NN construction and prediction, which provides a Monte Carlo estimation of a forecasting confidence interval. Within this approach, a large number of LSTM-NNs are constructed by resampling time-series sequences that were obtained from the early stage quantum evolution given by numerically exact multilayer multiconfigurational time-dependent Hartree method. The built LSTM-NN ensemble is used for the reliable propagation of the long-time quantum dynamics, and the simulated result is highly consistent with the exact evolution. The forecasting uncertainty that partially reflects the reliability of the LSTM-NN prediction is also given. This demonstrates the bootstrap-based LSTM-NN approach is a practical and powerful tool to propagate the long-time quantum dynamics of open systems with high accuracy and low computational cost.

18 citations

Journal ArticleDOI
TL;DR: In this paper, a nonparametric machine learning algorithm (kernel ridge regression as a representative of the kernel methods) is employed to study the quantum dissipative dynamics of the widely-used spin-boson model.
Abstract: The future forecasting ability of machine learning (ML) makes ML a promising tool for predicting long-time quantum dissipative dynamics of open systems. In this Article, we employ nonparametric machine learning algorithm (kernel ridge regression as a representative of the kernel methods) to study the quantum dissipative dynamics of the widely-used spin-boson model. Our ML model takes short-time dynamics as an input and is used for fast propagation of the long-time dynamics, greatly reducing the computational effort in comparison with the traditional approaches. Presented results show that the ML model performs well in both symmetric and asymmetric spin-boson models. Our approach is not limited to spin-boson model and can be extended to complex systems.

13 citations

Journal ArticleDOI
TL;DR: The QNODE as mentioned in this paper is a latent neural ordinary differential equation (ODE) trained on expectation values of closed and open quantum systems dynamics, which can learn to generate such measurement data and extrapolate outside of its training region that satisfies the von Neumann and time-local Lindblad master equations in an unsupervised way.
Abstract: The core objective of machine-assisted scientific discovery is to learn physical laws from experimental data without prior knowledge of the systems in question. In the area of quantum physics, making progress towards these goals is significantly more challenging due to the curse of dimensionality as well as the counterintuitive nature of quantum mechanics. Here we present the QNODE, a latent neural ordinary differential equation (ODE) trained on expectation values of closed and open-quantum-systems dynamics. It can learn to generate such measurement data and extrapolate outside of its training region that satisfies the von Neumann and time-local Lindblad master equations for closed and open quantum systems, respectively, in an unsupervised means. Furthermore, the QNODE rediscovers quantum-mechanical laws such as the Heisenberg's uncertainty principle in a data-driven way, without any constraint or guidance. Additionally, we show that trajectories that are generated from the QNODE that are close in its latent space have similar quantum dynamics while preserving the physics of the training system.

13 citations

Journal ArticleDOI
TL;DR: In this article , an AI-QD approach using AI to directly predict QD as a function of time and other parameters such as temperature, reorganization energy, etc., completely circumventing the need of recursive step-wise dynamics propagation in contrast to the traditional QD and alternative, recursive AI-based QD approaches.
Abstract: Exploring excitation energy transfer (EET) in light-harvesting complexes (LHCs) is essential for understanding the natural processes and design of highly-efficient photovoltaic devices. LHCs are open systems, where quantum effects may play a crucial role for almost perfect utilization of solar energy. Simulation of energy transfer with inclusion of quantum effects can be done within the framework of dissipative quantum dynamics (QD), which are computationally expensive. Thus, artificial intelligence (AI) offers itself as a tool for reducing the computational cost. Here we suggest AI-QD approach using AI to directly predict QD as a function of time and other parameters such as temperature, reorganization energy, etc., completely circumventing the need of recursive step-wise dynamics propagation in contrast to the traditional QD and alternative, recursive AI-based QD approaches. Our trajectory-learning AI-QD approach is able to predict the correct asymptotic behavior of QD at infinite time. We demonstrate AI-QD on seven-sites Fenna-Matthews-Olson (FMO) complex.

11 citations

References
More filters
Journal ArticleDOI
TL;DR: A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.
Abstract: Learning to store information over extended time intervals by recurrent backpropagation takes a very long time, mostly because of insufficient, decaying error backflow. We briefly review Hochreiter's (1991) analysis of this problem, then address it by introducing a novel, efficient, gradient based method called long short-term memory (LSTM). Truncating the gradient where this does not do harm, LSTM can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units. Multiplicative gate units learn to open and close access to the constant error flow. LSTM is local in space and time; its computational complexity per time step and weight is O. 1. Our experiments with artificial data involve local, distributed, real-valued, and noisy pattern representations. In comparisons with real-time recurrent learning, back propagation through time, recurrent cascade correlation, Elman nets, and neural sequence chunking, LSTM leads to many more successful runs, and learns much faster. LSTM also solves complex, artificial long-time-lag tasks that have never been solved by previous recurrent network algorithms.

72,897 citations

Journal ArticleDOI
28 May 2015-Nature
TL;DR: Deep learning is making major advances in solving problems that have resisted the best attempts of the artificial intelligence community for many years, and will have many more successes in the near future because it requires very little engineering by hand and can easily take advantage of increases in the amount of available computation and data.
Abstract: Deep learning allows computational models that are composed of multiple processing layers to learn representations of data with multiple levels of abstraction. These methods have dramatically improved the state-of-the-art in speech recognition, visual object recognition, object detection and many other domains such as drug discovery and genomics. Deep learning discovers intricate structure in large data sets by using the backpropagation algorithm to indicate how a machine should change its internal parameters that are used to compute the representation in each layer from the representation in the previous layer. Deep convolutional nets have brought about breakthroughs in processing images, video, speech and audio, whereas recurrent nets have shone light on sequential data such as text and speech.

46,982 citations

Journal ArticleDOI
01 Jan 1998
TL;DR: In this article, a graph transformer network (GTN) is proposed for handwritten character recognition, which can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters.
Abstract: Multilayer neural networks trained with the back-propagation algorithm constitute the best example of a successful gradient based learning technique. Given an appropriate network architecture, gradient-based learning algorithms can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters, with minimal preprocessing. This paper reviews various methods applied to handwritten character recognition and compares them on a standard handwritten digit recognition task. Convolutional neural networks, which are specifically designed to deal with the variability of 2D shapes, are shown to outperform all other techniques. Real-life document recognition systems are composed of multiple modules including field extraction, segmentation recognition, and language modeling. A new learning paradigm, called graph transformer networks (GTN), allows such multimodule systems to be trained globally using gradient-based methods so as to minimize an overall performance measure. Two systems for online handwriting recognition are described. Experiments demonstrate the advantage of global training, and the flexibility of graph transformer networks. A graph transformer network for reading a bank cheque is also described. It uses convolutional neural network character recognizers combined with global training techniques to provide record accuracy on business and personal cheques. It is deployed commercially and reads several million cheques per day.

42,067 citations

Proceedings ArticleDOI
Yoon Kim1
25 Aug 2014
TL;DR: The CNN models discussed herein improve upon the state of the art on 4 out of 7 tasks, which include sentiment analysis and question classification, and are proposed to allow for the use of both task-specific and static vectors.
Abstract: We report on a series of experiments with convolutional neural networks (CNN) trained on top of pre-trained word vectors for sentence-level classification tasks. We show that a simple CNN with little hyperparameter tuning and static vectors achieves excellent results on multiple benchmarks. Learning task-specific vectors through fine-tuning offers further gains in performance. We additionally propose a simple modification to the architecture to allow for the use of both task-specific and static vectors. The CNN models discussed herein improve upon the state of the art on 4 out of 7 tasks, which include sentiment analysis and question classification.

9,776 citations

Proceedings ArticleDOI
07 Dec 2015
TL;DR: The learned features, namely C3D (Convolutional 3D), with a simple linear classifier outperform state-of-the-art methods on 4 different benchmarks and are comparable with current best methods on the other 2 benchmarks.
Abstract: We propose a simple, yet effective approach for spatiotemporal feature learning using deep 3-dimensional convolutional networks (3D ConvNets) trained on a large scale supervised video dataset. Our findings are three-fold: 1) 3D ConvNets are more suitable for spatiotemporal feature learning compared to 2D ConvNets, 2) A homogeneous architecture with small 3x3x3 convolution kernels in all layers is among the best performing architectures for 3D ConvNets, and 3) Our learned features, namely C3D (Convolutional 3D), with a simple linear classifier outperform state-of-the-art methods on 4 different benchmarks and are comparable with current best methods on the other 2 benchmarks. In addition, the features are compact: achieving 52.8% accuracy on UCF101 dataset with only 10 dimensions and also very efficient to compute due to the fast inference of ConvNets. Finally, they are conceptually very simple and easy to train and use.

7,091 citations