Active learning algorithm for computational physics - Physical ... › publication › fulltext › Active-lea... › publication › fulltext › Active-lea...by J Yao · 2020 · Cited by 6 · Related articlesMar 10, 2020 — space is already time-demanding. In ... is
Active learning algorithm for computational physics Juan Yao ,1 Yadong Wu,2 Jahyun Koo,3 Binghai Yan,3 and Hui Zhai2,1,* 1
Center for Quantum Computing, Peng Cheng Laboratory, Shenzhen 518005, China 2 Institute for Advanced Study, Tsinghua University, Beijing 100084, China 3 Department of Condensed Matter Physics, Weizmann Institute of Science, Rehovot 76100, Israel (Received 16 May 2019; accepted 21 February 2020; published 10 March 2020) In large-scale computations of physical problems, one often encounters the situation of having to determine a multidimensional function, which can be numerically costly when computing each point in this multidimensional space is already time-demanding. In the work, we propose that the active learning algorithm can speed up such calculations. The basic idea is to fit a multidimensional function by neural networks, and the key point is to make the query of labeled data more economical by using a strategy called “query by committee.” We present the general protocol of this fitting scheme, as well as the procedure of how to further compute physical observables with the fitted functions. We show that this method can work well with two examples, which are the quantum three-body problem in atomic physics and the anomalous Hall conductivity in condensed matter physics, respectively. In these examples, we show that one reaches an accuracy of a few percent error in computing physical observables, all the while using fewer than 10% of total data points compared with uniform sampling. With these two examples, we also visualize that by using the active learning algorithm, the required amount of data points are added mostly in the regime where the function varies most rapidly, which explains the mechanism for the efficiency of the algorithm. We expect broad applications of our method to various kinds of computational physical problems. DOI: 10.1103/PhysRevResearch.2.013287
I. BACKGROUND
Neural network (NN) based supervised learning methods have nowadays found broad applications in studying quantum physics in condensed matter materials and atomic, molecular and optical systems [1,2]. On the theoretical side, applications include finding orders and topological invariants in quantum phases [3–9], generating variational wave functions for quantum many-body states [10–14] and speeding up quantum Monte Carlo sampling [15,16]. On the experimental side, these methods can help optimizing experimental protocols [17,18] and analyzing experimental data [19–21]. Usually the supervised learning scheme requires a huge set of labeled data. However, in many physical applications, labeling data can be quite expensive in terms of computational resources. For instance, performing computation or experiments repeatedly can be time- and resources-demanding. Therefore, in many cases, labeled data are not abundant, which is a challenge that has prevented many applications. The active learning is a scheme to solve this problem [22]. It starts from training a NN with a small initial data set, and then actively queries the labeled data based on the prediction of the NN and iteratively improves the performance of the NN
*
[email protected]
Published by the American Physical Society under the terms of the Creative Commons Attribution 4.0 International license. Further distribution of this work must maintain attribution to the author(s) and the published article’s title, journal citation, and DOI. 2643-1564/2020/2(1)/013287(6)
013287-1
until the goal of the task is reached. With this approach, sampling the large parameter space can be made more efficiently, and the demand of labeled data is usually much less than normal supervised learning methods. Recently a few works have applied the active learning algorithm to determination of the interatomic potentials in quantum materials [23–25] and to optimal control in quantum experiments [26,27]. In these situations, labeled data have to be obtained either by ab initio calculation or by repeating experiments, which are both time consuming. In this work, we focus on a class of general and common task in computational physics that is to numerically determine a multidimensional function, say, F (α1 , α2 , . . . , αn ), where αi are parameters. Supposing that we uniformly discretize each parameter into L points, there are then L n data points in total that need to be calculated. In many cases, calculation of each point already takes quite some time, and thus the total computational cost can be massive. Nevertheless, for most functions, there are regions where the function varies smoothly and regions where the function varies rapidly. Ideally, the goal is to sample more points in the steep regions and fewer points in the smooth regions in order to efficiently obtain a good fitting in the entire parameter space. However, it seems to be a paradox because one does not know which regions the function varies more rapidly prior to computing the function. Here we show that this goal can be achieved by using the active learning algorithm and the “query by committee” strategy [28] where dat