| Literature DB >> 33286901 |
Iosif Meyerov1, Evgeny Kozinov1, Alexey Liniov1, Valentin Volokitin1, Igor Yusipov1,2, Mikhail Ivanchenko1,2, Sergey Denisov1,3.
Abstract
With their constantly increasing peak performance and memory capacity, modern supercomputers offer new perspectives on numerical studies of open many-body quantum systems. These systems are often modeled by using Markovian quantum master equations describing the evolution of the system density operators. In this paper, we address master equations of the Lindblad form, which are a popular theoretical tools in quantum optics, cavity quantum electrodynamics, and optomechanics. By using the generalized Gell-Mann matrices as a basis, any Lindblad equation can be transformed into a system of ordinary differential equations with real coefficients. Recently, we presented an implementation of the transformation with the computational complexity, scaling as O(N5logN) for dense Lindbaldians and O(N3logN) for sparse ones. However, infeasible memory costs remains a serious obstacle on the way to large models. Here, we present a parallel cluster-based implementation of the algorithm and demonstrate that it allows us to integrate a sparse Lindbladian model of the dimension N=2000 and a dense random Lindbladian model of the dimension N=200 by using 25 nodes with 64 GB RAM per node.Entities:
Keywords: Lindblad equation; MPI; open quantum systems; parallel computing; performance optimization
Year: 2020 PMID: 33286901 PMCID: PMC7597275 DOI: 10.3390/e22101133
Source DB: PubMed Journal: Entropy (Basel) ISSN: 1099-4300 Impact factor: 2.524
The algorithm.
| Step | Substep (the Sequential Algorithm) | Parallelization |
|---|---|---|
|
| 1.1. Read the initial data from configuration files. | Sequential step, all the operations are performed on every node of a cluster. |
|
| This step is parallelized, computation time and memory costs are distributed among cluster nodes via Message Passing Interface (MPI). | |
| 2.1. Compute the coefficients | Step 2.1. ( | |
|
| 2.2. Compute the coefficients | Step 2.2. ( |
| 2.3. Compute the coefficients | Parallel steps 2.3, 2.4, and 2.5 are sketched in | |
| 2.6. Compute the initial value | Step 2.6 is not resource demanding and is realized on each cluster node independently. | |
|
| 3.1. Integrate the ODE ( | This step is parallelized via MPI (among cluster nodes), OpenMP (among CPU cores on every node), and SIMD instructions (on every CPU core). |
|
| 4.1. Save the results. | Here we gather results from computational nodes, save the results, and finalize MPI. |
Figure 1The parallel data preparation pipeline for the dimer model. On Panel A, we sketch distribution of nonzero elements of matrices S, J, and D, forming basis , Equations (3)–(5), respectively. Panel B depicts locations of nonzero elements in tensors f and d, Equation (11), which are not stored in memory but computed on the fly, during the Data Preparation step. Panels C, D, and E show how matrices Q, Equation (13), K, Equation (14), and R, Equation (15), are computed in parallel by two MPI processes (steps 2.3–2.5 in Table 1).
Figure 2Memory consumption per one node of the Data Preparation step for the sparse model.
Figure 3Memory consumption per one node (left) and computation time (right) of the Data Preparation step for the dense model.
Figure 4Computation time of ODE Integration step for the sparse (left) and dense (right) models. Numerical integration was performed for one period of modulation T, with 20,000 steps per period.
Figure 5Histogram of the complex eigenvalues, , , of a single realization of a random Lindbladian (see Section 3) for . The shown area was resolved with a grid of cells; the number of eigenvalues in every cell was normalized by the cell area. Altogether, eigenvalues are presented (except ). The bright thick line corresponds to the universal spectral boundary derived analytically in [17].