Abstract
Machine Learning methods are emerging as faster and efficient alternatives to numerical simulation techniques. The field of Scientific Computing has started adopting these data-driven approaches to faithfully model physical phenomena using scattered, noisy observations from coarse-grained grid-based simulations. In this paper, we investigate data-driven modelling of Bose-Einstein Condensates (BECs). In particular, we use Gaussian Processes (GPs) to model the ground state wave function of BECs as a function of scattering parameters from the dimensionless Gross Pitaveskii Equation (GPE). Experimental results illustrate the ability of GPs to accurately reproduce ground state wave functions using a limited number of data points from simulations. Consistent performance across different configurations of BECs, namely Scalar and Vectorial BECs generated under different potentials, including harmonic, double well and optical lattice potentials pronounces the versatility of our method. Comparison with existing data-driven models indicates that our model achieves similar accuracy with only a small fraction (\(\frac{1}{50}\)th) of data points used by existing methods, in addition to modelling uncertainty from data. When used as a simulator post-training, our model generates ground state wave functions \(36 \times \) faster than Trotter Suzuki, a numerical approximation technique that uses Imaginary time evolution. Our method is quite general; with minor changes it can be applied to similar quantum many-body problems.
Similar content being viewed by others
Introduction
At ultra-cold temperatures, when a gas is dilute enough such that the interactions are weak, the atoms crash into the ground state, leading to an exotic state of matter called Bose-Einstein Condensates (BECs)1. BECs are in general governed by the mathematical model described by Gross-Pitaevskii equation (GPE), a variant of Nonlinear Schrödinger equation (NLSE). In GPE, there are two variants, namely nonlinear interaction parameter (coupling strength) and the trapping potential and both could be spatio-temporal in general. GPE in its most general form cannot be solved analytically. It can be solved analytically2 only for specific choices of coupling strengths and trapping potentials. In other cases, one has to resort to numerical techniques3 like Crank-Nicholson Scheme, Trotter-Suzukii Approximation, Finite Element Analysis, etc.
Numerical techniques rely on numerical differentiation which is the finite difference approximation of derivatives using values of the original function evaluated at some sample points. Numerical approximations of derivatives are inherently ill-conditioned and unstable due to introduction of truncation and round-off errors4. On the other hand, Automatic Differentiation (AD) also called algorithmic differentiation is a family of techniques for efficiently and accurately evaluating derivatives of numeric functions expressed as computer programs5. AD exploits the fact that every computer program executes a sequence of elementary arithmetic operations and elementary functions. By applying the chain rule repeatedly to these operations, derivatives of arbitrary order can be computed automatically and accurately to working precision. AD is both efficient and numerically stable compared to Numerical Differentiation. AD has become the beating heart of Modern Machine Learning, leading to a new paradigm of programming called Differentiable Programming.
In recent years, Scientific Computing community has started adopting data-driven approaches from Machine Learning, as faster and efficient alternatives to numerical simulation techniques. This new regime called Data-driven Scientific Computing has shown great promise already by faithfully modelling physical phenomena using scattered noisy observations from coarse-grained grid-based numerical simulations. The learned models are capable of predicting the dynamics of the system under study in a fixed number of CPU cycles irrespective of the dimensions of the grid. Unlike numerical techniques that rely on unstable numerical gradients, Machine Learning models depend on precise and stable automatic differentiation.
Advances in Machine Learning has led to incredible breakthroughs in areas such as Computer Vision6, Natural Language Processing7, Computational Biology8, etc, demonstrating super-human abilities9,10,11,12 in certain tasks. These dramatic improvements are reflected in computational sciences that make use of Machine Learning. Artificial Neural Networks (ANNs), the focal point of Modern Machine Learning or Deep Learning, are powerful and versatile models that can automatically learn representations from data. ANNs were primarily used in Physical Sciences, to learn suitable order parameters and detect the phase transition, without any prior domain knowledge13,14,15,16. Greitemann et al.17 used a kernel-based learning method to learn the phases in magnetic materials to identify complex order parameters. In the context of nonlinear dynamical systems, Jaegar et al.18 employed ANNs to predict the trajectories of a chaotic time-series improving the accuracy by a factor of 2400 over previous techniques. ANNs were used as a variational representation of quantum states in quantum-physical many-body problems to cope up with the exponential complexity of the many-body wave function19. Nomura et al.20 employed ANNs to develop a machine learning method for constructing accurate ground-state wave functions of strongly interacting and entangled quantum spin systems as well as fermionic models on lattices. A restricted Boltzmann machine algorithm in the form of an ANN, combined with a conventional variational Monte Carlo method turned out to be a highly accurate quantum many-body solver capable of accurately predicting ground-state wave functions of strongly interacting and entangled quantum spin systems as well as fermionic models on lattices21. ANNs have been proven to be capable of representing large-scale quantum states, including the ground states of many-body Hamiltonians and states generated by quantum dynamics22,23,24,25.
Despite being robust and flexible, ANNs suffer from a number of shortcomings. ANNs are data-hungry and hence they do not play well with small data. They are over-confident in predictions and are susceptible to adversarial inputs. Learning in ANNs is based on Maximum Likelihood Estimation which leads to point estimates of learnable parameters instead of probability distributions. Gaussian Processes (GPs) on the other hand, are flexible probabilistic models that can learn probability distributions over functional mapping of inputs to outputs. They are quite efficient in learning hidden dynamics from small data. Being a Bayesian method, GP inherently models uncertainty from data. Learning results in a posterior distribution from which we could draw new samples.
Even though Machine Learning approach has been employed in modelling BECs earlier by Liang et al.26, they exploited Convolutional Neural Network architecture(CNN) to model the ground state wave functions of BECs in different platforms. Liang et al.26 showed that it is possible to approximate the wave function using a neural network but do not guarantee any kind of improvement over conventional numerical techniques27. Our investigation bridges this gap by using a Gaussian Process as a cheaper, powerful and efficient alternative to neural networks in the context of modelling ground state wave functions in BECs.
In this paper, our goal is to examine whether Gaussian Processes could be a viable surrogate model for data-driven emulation of one-dimensional Scalar and Vectorial BECs. Our validation of GP is centred around the following aspects:
-
1.
Correctness: Can GP accurately model the ground state of BECs?
-
2.
Versatility: Can GP adapt to different settings of BECs?
-
3.
Data efficiency: How many data points are necessary to model a wave function?
-
4.
Compute efficiency: How do they fare against numerical techniques when used as a simulator, after training?
To answer these questions, we studied the performance of Gaussian Processes on different settings of BECs. We employ Trotter-suzuki approximation, a numerical technique for simulating the ground state wave functions, \(\psi \). We run simulations by setting up a one-dimensional grid and vary the coupling strength g, the parameter that controls the interaction between the atoms besides varying the trapping potential. Similarly for two-component BECs, we run simulations by varying interaction parameters \(\{g_{11}, g_{12}, g_{22}\}\) which results in two wave functions \(\{\psi _1, \psi _2\}\). We model the simulated wave functions using a GP with Radial Basis Function (RBF) kernel28. The GP models the wave function as a function of space x and coupling strength g. The results reveal the versatility of GP in modelling different kinds of wave functions, efficiently and accurately, with uncertainty estimates. In addition to modelling uncertainty from data, our method performs better than Liang et al.26 in terms of efficiency and model complexity. Our model being simpler in terms of model complexity, uses just a small fraction (\(\frac{1}{50}\)th) of the data points used by Liang et al. to achieve similar accuracy. Furthermore, comparing the efficiency of our method in predicting wave functions, with Trotter-suzuki approximation, we find that our method performs \(36 \times \) faster.
The paper is organised as follows: “Background” section lays the foundation for the rest of the paper. It starts with a brief introduction to Bose-Einstein Condensates in “Bose-Einstein Condensates” section, Trotter-Suzuki Approximation in “Trotter-Suzuki approximation” section, and a detailed description of Gaussian Processes in “Gaussian Processes”. “Methods” section provides the reader with the tools necessary to replicate our experiments. “Experiments” section includes a report on all our experiments using Gaussian Processes as a surrogate model for different settings of BECs. “Discussion” section discusses the implications of our experimental results and summarizes possible avenues for further research.
Background
Bose-Einstein Condensates
At temperatures close to absolute zero, majority of the bosons in a gas crash into the ground state of the system creating an exotic state of matter known as Bose-Einstein Condensates (BECs). The dynamics of BECs is described by Gross-Pitaevskii29 Eq. (1) which belongs to the family of Variable Coefficient Nonlinear Schrödinger equations, given by
where \(\psi \) denotes the wave function(or order parameter) of BECs, m is mass the atoms of the condensate, g is the inter-atomic interaction, \(V_{ext}\) is the trapping potential and \(|\psi |^{2}\) is the atomic density.
A two component BEC known as Vector BEC is endowed with inter-atomic and intra atomic interaction in addition to the trap and hence its density can be manipulated with more freedom and is governed by Coupled Gross-Pitaveskii equation3 of the form Eqs. (2) and (3),
Trotter-Suzuki approximation
The behaviour of any physical system can be studied by solving the partial differential equations (PDEs) which represent the dynamics of that physical phenomenon. In practice, most PDEs with any real application are nonlinear in nature and are hard to solve analytically. This is particularly true for complex dynamical systems which are quite difficult to solve and necessitate high computational resources to arrive at highly accurate solutions. Trotter-suzuki decomposition implemented by Wittek and Cucchietti30, exploits optimized kernels to solve Gross-Pitaevskii equation of a free particle. The exponential operators in PDEs are notoriously hard to approximate. Trotter Suzuki decomposes the Hamiltonian into sum of diagonal matrices which eases the task of computing the exponential. The evolution operator is calculated using the Trotter-Suzuki approximation. Given an Hamiltonian as a sum of hermitian operators, for instance \(H=H_1+H_2+H_3\), the evolution is approximated as30
Gaussian Processes
A Gaussian Process (GP)31 are probabilistic machine learning models that can be viewed as a distribution over a set of functions. A multi-variate Gaussian distribution is a generalization of the one-dimensional gaussian distribution. Considering functions as infinite-dimensional vectors, we can say that a GP is an infinite dimensional generalization of a multi-variate Gaussian distribution. A sample from GP is a realization from the set of admissible functions. The set of admissible functions is defined by the mean vector and the covariance matrix of the GP which are in turn generated by a mean function and a covariance function. This space of functions is constrained via the mean and covariance functions by making assumptions about the smoothness, stationarity and sparsity of the set of functions. These assumptions result in a mean and covariance that jointly represent the GP prior. We use Bayes’ rule to constrain the prior on observed data. Given a set of points in the input space \(\{x_1, x_2, .. x_n\}\) and the function f evaluated at those points \(\{f^{1}_{e}, f^{2}_{e}, ..., f^{n}_{e} \}\), we can formally define a GP as:
Definition
p(f) is a Gaussian Process if for any subset \(\{x_1, x_2, ... x_n\} \subset {\mathcal{X}}\), the marginal distribution over the subset \(p(f_{e})\) has a multivariate Gaussian distribution.
Consider a Multivariate Gaussian Distribution given by,
where \(\mu \in {\mathbb {R}}^k\) is a k-dimensional zero vector and \(\Sigma \in {\mathbb {R}}^{k \times k}\) is the covariance matrix that captures the correlation between dimensions \(\{X_1, X_2, .. X_{k}\}\). Let us sample 4 points \(\{z_1, z_2, z_3, z_4 \}\) from the distribution and plot it sequentially as shown in Fig. 1a. From Fig. 1a, we can make two inferences. (1) Points closer to each other on the x-axis behave similarly on the y-axis (2) Points farther apart from each other behave differently on the y-axis. That is, closer points are highly correlated with each other while father points are not. This intuition leads to the understanding that by connecting these points sequentially, we can realize smooth functions like the ones shown in Fig. 1b. The functions we realized by connecting the points sequentially can be considered as samples from a Gaussian Process by a zero-mean vector \(\mu \) and the hitherto unknown covariance matrix \(\Sigma \).
Yet another inference that can be made from Fig. 1a is that the nature of the functions realized are dependent on the covariance matrix \(\Sigma \). By manipulating the covariance matrix \(\Sigma \), one can manipulate the kind of functions realized from k-dimensional samples \(\{z_i\}\). The mechanism to manipulate the covariance matrix is through an algebraic function named Covariance function or kernel. The kernel \(k(x, x') \rightarrow {\mathcal{R}}\) takes two points from the input space \({\mathcal{X}}\) as inputs and returns a scalar value \(c \in [0, 1]\). This value represents the similarity between the two input points. The covariance matrix \(\Sigma \) is constructed by calculating the similarity between k equally spaced points in the input space \(\{x_1, ... x_k\}\). By controlling the kernel function, we control the covariance matrix that governs the GP, thereby controlling the nature of functions generated. The kernel is usually parameterized by one or more tunable variables which act as knobs for controlling the shape of the functions that GP generates.
Let us assume for didactic purposes that the x and y are scalars and the function f is given by \(f : {\mathcal{X}} \rightarrow R\). We consider Radial Basis Function (RBF) parameterized by length scale \(k_l\) as the kernel.
Length scale \(k_l\) controls the order of the functions generated as evident from Fig. 1b.
Inference
Inference in GP is a two-step process : Kernel Parameter Search and Posterior Estimation. In kernel parameter search, we learn the kernel parameter values that fit the data well. The optimal value of kernel parameters are estimated using Maximum Likelihood Estimation (MLE) by minimizing marginal negative log likelihood32 given by
where N is the number of data points, (X, y) represent the data points.
Then, we estimate GP posterior conditioned on data i.e. mean vector and covariance matrix conditioned on data. The definition of GP states that the joint distribution of observed data y and predictions \(f_*\) has a multivariate Gaussian distribution. Given data X and new input \(X_{*}\), we can write the joint distribution as,
Using the Multivariate Gaussian Theorem33, we arrive at the parameters of the posterior,
Motivation
In our experiments, we consider the ground state wave function of one-dimensional BECs27. The simulation takes as input multiple parameters including grid parameters like radius, length, potential function (\(V_{ext}\)), coupling strength (g),etc., and generates a one-dimensional wave function using Imaginary Time Evolution method. The time period to complete a simulation depends on the dimensions and size of the grid. The promise of Data-driven Scientific computing is the ability to model any process without closed-form solutions using experimental data and building a predictive model that can provide predictions at any point in the grid with absolute guarantee. As discussed in the Introduction, a data-driven model trained on scattered data points from simulations is capable of predicting the wave function at any point within the grid in a few fixed number of CPU cycles irrespective of grid dimensions or size.
Gaussian Processes model the wave function as a probability distribution over functions supported by a kernel with optimal parameters. Being a Bayesian model, GP is data-efficient and makes uncertainty-aware predictions. The ability to model uncertainty is especially important in modelling physical phenomenon from a few scattered noisy observations. Bayesian methods rely on Bayesian Inference which estimates posterior over unknowns in contrast to non-probabilistic models which result in point estimates of unknowns. As a consequence, it is possible to generate entire wave functions from the GP that are subject to constraints on the input parameters like coupling strength, omega(Rabi Coupling), etc.
Methods
We use trotter-suzuki-mpi34, a massively parallel implementation of the Trotter-Suzuki approximation to simulate the evolution of Bose-Einstein Condensates. The simulation setup consists of a one-dimensional grid which represents the physical system. This discretized space is defined by a lattice of 512 nodes within a physical length of 24 units. The physics of BECs is described by the Hamiltonian which represents the GPE equation. The Hamiltonian requires a grid and a trapping potential. The trapping potential is defined as a function which takes x and y as arguments and returns a scalar value as output. In this case, we consider a harmonic trapping potential given by \(\frac{x^2}{2}\). The state of the system is initialized in gaussian form. Imaginary time evolution evolves the system state using the dynamics defined by the Hamiltonian, in \(10^{3}\) iterations with time step being \(\Delta _t = 10^{-4}\) for each iteration. In order to collect data, we run M simulations by varying the value of the interaction parameter g. We use sklearn’s Gaussian Process API35 with Radial Basis Function (RBF) kernel as a surrogate to model the wave function \(\psi \) as a simpler but continuous function of space x and the interaction parameter g.
The prior on the groundstate wave function of BEC is given by
N data points of the form \((x, g, \psi )\) are sampled randomly from M simulation outcomes. Throughout our experiments, we use the RBF kernel given in Eq. (5).
The hyper-parameters \(\theta \) of the kernel are tuned using Maximum Likelihood Estimation which essentially amounts to finding the parameters that minimize the expression in Eq. (6), given a list of data points.
The RBF kernel consists of the following parameters: length scale \(k_{l}\) and variance \(k_{\sigma }\). \(\psi _{posterior}\) is estimated by conditioning \(\psi _{prior}\) on the sampled data points. The expression for \(\psi _{posterior}\) reduces to a mathematically tractable form given by the mean vector and covariance matrix presented in equation 8. Inference consists of calculating the marginal mean \(\mu _{2|1}\) and covariance \(\Sigma _{2|1}\) by substituting the training points X and new test points \(X_*\) into the analytical forms presented in equation 8. The marginal mean \(\mu _{2|1}\) constitutes the predicted wave function. The diagonal elements of the covariance matrix form the variance \(\sigma \) on each prediction. Together, they make the GP posterior parameterized by \([\mu _{2|1}, \Sigma _{2|1}]\). The fidelity of predicted wave function is measured by calculating Mean Squared Error (MSE) metric against the ground truth data. Figure 2 portrays the results of an illustrative experiment on one-dimensional BECs with a harmonic trapping potential.
The experimental setup described above can be extended to different settings of simulation and data-driven approximation, to include two-component BECs and different kinds of trapping potentials. A detailed report of the experiments we conducted is presented in the next section.
Experiments
All the experiments to be covered in this section will follow along similar lines as the illustrative experiment discussed in the previous section: Simulation, Data-driven Approximation and Evaluation. In the previous experiment, we ran 300 simulations shown in Fig. 2a for equally spaced values of g ranging from 1 to 100. We randomly sampled 500 samples of the form \((x, g, \psi )\) from the simulation results, which we used to fit the GP. Evaluation leads to a MSE score of \(1.23 \times 10^{-7}\) which indicates that the trained GP can accurately predict ground state wave functions for 100 different values of the interaction parameter g within the range (1, 100).
To follow up, we ask the question, How many data points are necessary to build a reliable predictive model of ground state wave function? In order to answer this question, we vary the number of training samples and observe the effect it has on MSE score of GP trained on those samples. We reuse the experimental setup from the previous experiment and vary the number of data points N sampled from \(M=300\) simulations. The results are presented in Fig. 3. Figure 3 depicts the relationship between variance and number of samples N and it shows the rate of decrease of MSE w.r.t N. The MSE curve shows a trend of saturation close to 0 as the number of samples are increased as shown in Table 1. Decrease in variance and MSE is evidence of confidence and accuracy in prediction respectively.
The next experiment tests the versatility of GP by modelling the ground states of BECs for different choices of trapping potentials. We consider Harmonic , Double Well and Optical Lattice potentials in Fig. 4 . We restrict the range of g to (0, 2) in order to ensure the stability of generated wave function for different potentials. We run 100 simulations and collect 500 samples for each potential. We fit a GP for each potential using collected data points. Evaluating the models results in MSE scores of \(1.23 \times 10^{-7}\), \(1.57 \times 10^{-6}\) and \(4.30 \times 10^{-5}\), for harmonic, double well and optical lattice potentials3 respectively. The predictive wave functions for a value of \(g=1\) is plotted against the ground truth in Fig. 4a–c.
The ground state wave functions of two-component BECs are given by \(\psi _1\) and \(\psi _2\). The interaction strength is defined by parameters \(g_{11}\), \(g_{12}\) and \(g_{22}\). We model the system using a GP which learns the relationship between \((\psi _1, \psi _2)\) and the interaction parameters. We sample 500 data points from 200 simulations with values of interaction parameters lying within the range \((-1, 1)\). The choice of this range is dictated by the zone of stability of the system. Predictions and ground truth wave functions for different potentials conditioned on values of interaction parameters given by \(\{g_{11}=0.1, g_{12}=-0.1, g_{22}=0.1\}\), are presented in Fig. 4a–c. Evaluation on a test set within the range \(g \in (-1, 1)\) results in MSE in the range of \(10^{-4}\).
We reproduce an experiment conducted by Liang et al. 26 using GP-Emulation of rabi-coupled two-component BECs. Figure 4d shows the predicted ground state wave function of rabi-coupled two-component BECs Eqs. (9) and (10), conditioned on a cosine potential given by \(\frac{x^2}{2} + 24cos^2 x\) with interaction parameters \(\{g_{11}=103, g_{12}=100, g_{22}=97\}\). Finally, we use trained GPs as simulators to generate wave functions and compare their performance against Trotter-suzuki. We experiment with one-component and two-component BECs in 512 and 1024 grid spaces. The results tabulated in Table 2 indicate an average of \(36 \times \) speed up in wave generation with reasonable accuracy.
Discussion
CNN-based model employed by liang et al.26 learns a mapping from coupling strength g to wave function values at fixed points in space, a scalar to vector mapping. This results in a highly limited predictive model of wave functions. In contrast, our method models the wave function as a continuous function of space and coupling strength, resulting in a predictive model which is well-defined in all the points within the chosen interval. The Mean-Squared Error (MSE) presented by Liang et al.26, on an average ranges from \(10^{-4}\) to \(10^{-5}\). Our experimental results tabulated in Table 1 show a similar MSE score ranging from \(10^{-4}\) to \(10^{-7}\). We use a maximum of 1000 data points from simulations to achieve these results while Liang et al.26 uses data from 50000 simulations to train their CNN. Our method, only using a small fraction (\(\frac{1}{50}\)) of data points used by Liang et al.26, has demonstrated comparable accuracy in most cases and enhanced accuracy in others. Our experiments with various trapping potentials show that it is possible to model different kinds of ground states of BECs using a vanilla GP with an RBF kernel.
We compare the predictive power of our model with Trotter-suzuki Approximation. Trotter-suzuki generates ground state wave functions by setting up a grid and running time evolution given a set of initial conditions (configuration). This scales linearly with respect to the number of simulations. The advantage in using a trained GP as a predictive model is that multiple simulation runs can be batched together by converting different configurations \(\{(x, g_1), (x, g_2), ... (x, g_M)\}\) into an input tensor of form \([[x, x, ... x], [g_1, g_2, ... g_M]]\). This trick makes it possible to pose multiple simulations as one prediction step. This is the reason for the significant speed up in generation of wave function using GPs compared to Trotter-suzuki. The empirical results discussed thus far indicate that Gaussian Processes are not just viable but desirable surrogate models for data-driven emulation of Bose-Einstein Condensates.
There are a number of avenues to explore which are beyond the scope of this work. We briefly mention these in this section. Vanilla Gaussian Processes do not scale well to high-dimensional data. modelling higher-dimensional BECs requires a scalable variant of GP. GP inherently models uncertainty in data. It is possible to exploit this in order to train GPs even more efficiently using Active Learning. We used different instances of GP to model wave functions conditioned on different trapping potentials. Using a common Multi-task GP (MTGP)36 would increase the data-efficiency further as MTGP learns the common traits between wave functions generated in different settings.
Conclusion
Data-driven methods are replacing conventional numerical techniques in modelling physical phenomena in several fields of science. We explore data-driven emulation of BEC ground-state wave function using Gaussian Process as a surrogate model. Our empirical results show that Gaussian Processes in addition to modelling uncertainty from data, are able to surpass Artificial Neural Networks in terms of efficiency and complexity. Gaussian Processes when used as simulators post-training offer significant gain in speed when compared to Trotter-suzuki, a numerical approximation technique. Our experimental results indicate that GP is a desirable candidate for data-driven modelling of Bose-Einstein Condensates.
References
Ketterle, W. Experimental studies of Bose-Einstein condensation. Phys. Today 52, 30–35 (1999).
Radha, R. & Vinayagam, P. S. An analytical window into the world of ultracold atoms. Roman. Rep. Phys. 67, 89–142 (2015).
Bao, W. & Cai, Y. Mathematical theory and numerical methods for Bose-Einstein condensation. arXiv preprint arXiv:1212.5341 (2012).
Baydin, A. G., Pearlmutter, B. A., Radul, A. A. & Siskind, J. M. Automatic differentiation in machine learning: a survey. J. Mach. Learn. Res. 18, 5595–5637 (2017).
Rall, L. Automatic differentiation-technique and applications. Lecture Notes in Computer Science120 (1981).
Tan, M., Pang, R. & Le, Q. V. Efficientdet: scalable and efficient object detection. arXiv preprint arXiv:1911.09070 (2019).
Ng, N., Edunov, S. & Auli, M. Facebook artificial intelligence (2019).
Christiansen, E. M. et al. In silico labeling: predicting fluorescent labels in unlabeled images. Cell 173, 792–803 (2018).
Weyand, T., Kostrikov, I. & Philbin, J. Planet-photo geolocation with convolutional neural networks. In European Conference on Computer Vision, 37–55 (Springer, 2016).
Brown, N. & Sandholm, T. Superhuman ai for multiplayer poker. Science 365, 885–890 (2019).
Ibarz, B. et al. Reward learning from human preferences and demonstrations in atari. Adv. Neural Inf. Process. Syst. 8011–8023, (2018).
Borowiec, S. Alphago seals 4-1 victory over go grandmaster lee sedol. Guardian 15, (2016).
Carrasquilla, J. & Melko, R. G. Machine learning phases of matter. Nat. Phys. 13, 431–434 (2017).
Morningstar, A. & Melko, R. G. Deep learning the ising model near criticality. J. Mach. Learn. Res. 18, 5975–5991 (2017).
Tanaka, A. & Tomiya, A. Detection of phase transition via convolutional neural networks. J. Phys. Soc. Jpn. 86, 063001 (2017).
Zdeborová, L. Machine learning: new tool in the box. Nat. Phys. 13, 420–421 (2017).
Greitemann, J. et al. Identification of emergent constraints and hidden order in frustrated magnets using tensorial kernel methods of machine learning. Phys. Rev. B 100, 174408 (2019).
Jaeger, H. & Haas, H. Harnessing nonlinearity: predicting chaotic systems and saving energy in wireless communication. Science 304, 78–80 (2004).
Carleo, G. & Troyer, M. Solving the quantum many-body problem with artificial neural networks. Science 355, 602–606 (2017).
Nomura, Y., Darmawan, A. S., Yamaji, Y. & Imada, M. Restricted Boltzmann machine learning for solving strongly correlated quantum systems. Phys. Rev. B 96, 205152. https://doi.org/10.1103/PhysRevB.96.205152 (2017).
Gao, X. & Duan, L.-M. Efficient representation of quantum many-body states with deep neural networks. Nat. Commun. 8, 1–6 (2017).
Carleo, G., Nomura, Y. & Imada, M. Constructing exact representations of quantum many-body systems with deep neural networks. Nat. Commun. 9, 1–11 (2018).
Czischek, S., Gärttner, M. & Gasenzer, T. Quenches near ising quantum criticality as a challenge for artificial neural networks. Phys. Rev. B 98, 024311 (2018).
Schmitt, M. & Heyl, M. Quantum dynamics in transverse-field ising models from classical networks. SciPost Phys. 4, 013 (2018).
Fabiani, G. & Mentink, J. H. Investigating ultrafast quantum magnetism with machine learning. SciPost Phys. 7, 4. https://doi.org/10.21468/SciPostPhys.7.1.004 (2019).
Liang, X., Zhang, H., Liu, S., Li, Y. & Zhang, Y.-S. Generation of Bose-Einstein condensates’ ground state through machine learning. Sci. Rep. 8, 1–8 (2018).
Bao, W., Jaksch, D. & Markowich, P. A. Numerical solution of the gross-pitaevskii equation for Bose-Einstein condensation. J. Comput. Phys. 187, 318–342 (2003).
Vert, J.-P., Tsuda, K. & Schölkopf, B. A primer on kernel methods. Kernel Methods Comput. Biol. 47, 35–70 (2004).
Rogel-Salazar, J. The gross-pitaevskii equation and Bose-Einstein condensates. Eur. J. Phys. 34, 247 (2013).
Wittek, P. & Cucchietti, F. M. A second-order distributed Trotter–Suzuki solver with a hybrid CPU–GPU kernel. Comput. Phys. Commun. 184, 1165–1171 (2013).
MacKay, D. The humble gaussian distribution (2006). https://www.seas.harvard.edu/courses/cs281/papers/mackay-2006.pdf.
Rasmussen, C. E. Gaussian processes in machine learning. In Summer School on Machine Learning, 63–71 (Springer, 2003).
Murphy, K. P. Machine Learning: A Probabilistic Perspective (MIT Press, Cambridge, 2012).
Calderaro, L., Wittek, P. & Liu, D. Massively parallel Trotter–Suzuki solver. https://github.com/trotter-suzuki-mpi/trotter-suzuki-mpi (2017).
Buitinck, L. et al. API design for machine learning software: experiences from the scikit-learn project. ECML PKDD Workshop: Languages for Data Mining and Machine Learning 108–122, (2013).
Bonilla, E. V., Chai, K. M. & Williams, C. Multi-task gaussian process prediction. Adv. Neural Inf. Process. Syst. 153–160, (2008).
Acknowledgements
R.Radha wishes to acknowledge financial assistance received from Council of Scientific and Industrial Research (No. 03(1456)/19/EMR-II Dated: 05/08/2019), Government of India.
Author information
Authors and Affiliations
Contributions
S.R. and T.B. conceived the experiments, S.R. and T.B. conducted the experiments, M.S., R.R. and V.S. analysed the results. All authors reviewed the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Bakthavatchalam, T.A., Ramamoorthy, S., Sankarasubbu, M. et al. Bayesian Optimization of Bose-Einstein Condensates. Sci Rep 11, 5054 (2021). https://doi.org/10.1038/s41598-021-84336-0
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-021-84336-0
This article is cited by
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.