Self-correcting quantum memory with a boundary Adrian Hutter,1, ∗ James R. Wootton,1 Beat R¨othlisberger,1 and Daniel Loss1
arXiv:1206.0991v3 [quant-ph] 16 Nov 2012
1
Department of Physics, University of Basel, Klingelbergstrasse 82, CH-4056 Basel, Switzerland (Dated: November 19, 2012)
We study the two-dimensional toric code Hamiltonian with effective long-range interactions between its anyonic excitations induced by coupling the toric code to external fields. It has been shown that such interactions allow to increase the lifetime of the stored quantum information arbitrarily by making L, the linear size of the memory, larger [Phys. Rev. A 82 022305 (2010)]. We show that for these systems the choice of boundary conditions (open boundaries as opposed to periodic boundary conditions) is not a mere technicality; the influence of anyons produced at the boundaries becomes in fact dominant for large enough L. This influence can be both beneficial or detrimental. In particular, we study an effective Hamiltonian proposed in [Phys. Rev. B 83 115415 (2011)] that describes repulsion between anyons and anyon holes. For this system, we find a lifetime of the stored quantum information that grows exponentially in L2 for both periodic and open boundary conditions, though the exponent in the latter case is found to be less favourable. However, L is upper-bounded through the breakdown of the perturbative treatment of the underlying Hamiltonian. PACS numbers: 03.67.Lx, 03.67.Pp, 05.30.Pr, 75.10.Jm
I.
INTRODUCTION
An important open problem in quantum information concerns the feasibility of a self-correcting quantum memory. Finding a system that protects a quantum state from decoherence induced by a thermal bath, without the need for active monitoring and error-correction, proves much more difficult than in the classical case. On the most basic level, this is due to the fact that a classical bit only needs protection against logical X operations while a qubit needs protection against a logical X and Z. If a state is stored in a many-qubit system, a desirable feature is topological protection of the stored (qu)bit: we want a logical error X (or Z in the quantum case) to necessitate a number of single-qubit errors σx (or σz ) that scales with L, the linear size of the memory. The simplest model that energetically penalizes σx errors and offers topological protection of a stored classical bit is the 1D ferromagnetic Ising model. The simplest model that penalizes σx and σz errors and topologically protects a qubit is given by Kitaev’s 2D toric code Hamiltonian [1]. In fact, the latter can be mapped exactly to two independent copies of the former [2, 3]. Unfortunately, both of these systems are not thermally stable. Once a topological defect (a pair of domain walls in the 1D Ising model or a pair of anyons in the 2D toric code) has been created, it can spread and lead to a logical error without any further energy cost. The lifetime of a qubit stored in the degenerate ground states of the 2D toric code is thus for any finite temperature upper-bounded by a constant independent of L [3–5]. The 2D toric code can therefore not serve as a ‘quantum hard drive’. These difficulties can be overcome if the dimensionality of the systems is increased. In the 2D Ising model and the 4D toric code,
∗
[email protected]
any sequence of single-qubit Pauli operators that leads to a logical error has to surpass an energy barrier whose size scales with L. Since the number of error paths connecting two distinct ground states is exponentially large in L, these systems are thermally stable below some critical temperature Tc , meaning that the lifetime of the stored information grows exponentially with L [3, 6, 7]. Whether a similar degree of protection for a quantum state can be achieved in less than four dimensions is not clear. One can show that for every 2D local stabilizer Hamiltonian the height of the energy barrier separating orthogonal states stored in a degenerate ground state is upper bounded by a constant independent of L [8–10], ruling out the possibility of using such systems for the fault-tolerant storage of quantum information by selfcorrection. In principle, these no-go results leave two ways out: Either one abandons the locality of the terms in the Hamiltonian or one goes to dimension 3. Indeed, both of these routes have been followed in the recent literature. While the 3D toric code is not thermally stable, Haah showed in a recent breakthrough the existence of 3D Hamiltonians with local interactions that have no string-like logical operators [11]. Unlike anyons in the 2D toric code, defects cannot move further than a certain constant distance away without creating other defects. This property implies a logarithmically growing energy barrier between orthogonal ground states, leading one to expect a lifetime that grows polynomially with L [12]. However, the best known lower bound on the memory lifetime of Haah’s Hamiltonian is upper-bounded by a constant independent of L [13] and further improvement is not expected [14]. Furthermore, a 3D architecture may lead to practical difficulties when accessing the physical qubits for syndrome measurement and error correction. We therefore believe that the most promising route to follow in search for a realistic proposal for a quantum memory is to start from the 2D toric code Hamiltonian and add terms to it that
2 • can be physically motivated, and • lead to a memory lifetime that becomes arbitrarily large as L → ∞ . Long-range repulsive interactions (1/rα -potential with 0 ≤ α < 2) between the anyons lead to a logarithmicallygrowing self-consistent mean field gap for anyon creation, yielding a polynomially increasing lifetime [5], see Sec. III below. So rather than trying to find a Hamiltonian with a macroscopic energy barrier between orthogonal ground states, this approach seeks to suppress the anyon creation rate. The toric code Hamiltonian (involving local four -qubit couplings) with non-interacting anyons can be obtained as an effective Hamiltonian of the Kitaev honeycomb model, which involves nearest-neighbor two-qubit Ising couplings [15]. How an α = 0 interaction between the anyons can be obtained through such a honeycomb model coupled to electromagnetic modes has been studied in detail in [16], see Sec. IV below. In an alternative approach it was shown that coupling the toric code to a bosonic field leads to an effective gravitational potential between the anyonic defects [17]. Below some critical temperature, all anyons coalesce to a single point. However, the time the system needs to approach this metastable state and how to best perform error correction in this system have not been investigated so far. Self-correcting quantum memories are usually discussed with periodic boundary conditions, giving the toric code its name. This way, the complications that arise with the possibility of creating unpaired topological defects at the boundaries can be avoided. One expects that the influence of the boundaries becomes negligible if L becomes large enough, which is certainly correct for Hamiltonians with local interactions. However, here we study memories with long-range interactions between the anyons as proposed in [5, 16] and show that for these systems the influence of the boundary becomes in fact dominant for large enough L. It can be beneficial and detrimental. Specifically, unpaired anyons from the boundary lead to an effective bias for anyons from the bulk to move to the closest boundary, thus prolonging the time until error correction becomes ambiguous (see Sec. III). On the other hand, the ability to create unpaired anyons at the boundaries halves the energetic gap above the anyonic vacuum. This becomes especially relevant if this gap is so strong that the anyonic system is basically restricted to its ground state and first excited state (see Sec. IV). The paper is structured as follows. In Section II we discuss how error correction can be performed in the planar code (a toric code with open boundaries) in contact with a thermal environment. In Section III we study the influence of the boundaries for a Hamiltonian with spatially constant repulsion between the anyons, while in Section IV an effective Hamiltonian that describes repulsion between anyons and anyon holes is investigated.
II.
ERROR CORRECTION IN THE PLANAR CODE A.
The planar code
A self-correcting quantum memory is supposed to protect a quantum state from a thermal environment by means of its internal dynamics and without need for active error-correction. A single error correction step may be performed before the stored state is read out. We shall use here a version of the toric code first introduced in [18], which, contrary to what the name suggests, is not periodic but does have a boundary. We will refer to this as the planar code. Consider a grid with quadratic cells and physical qubits placed on the edges, as depicted in Figure 1. We call the four qubits around one unit cell a ‘plaquette’ and the four qubits around a vertex a ‘star’. We define plaquette operators Ap = (σz )⊗4 , where the tensor product runs over the four qubits around some plaquette p and star operators Bs = (σx )⊗4 , where the tensor product runs over the four qubits around some vertex s. Plaquette operators on the left and right boundary and star operators on the top an bottom boundary are tensor products of three Pauli operators only. All of these operators are commuting since they overlap at zero or two qubits. Let the space K0 ⊂ H = (C2 )⊗N (N is the total number of qubits) be defined as the space stabilized by all plaquette and star operators. That is, K0 is the space of all states |ψi such that for each three- or four-qubit plaquette or star operator Ap or Bs we have Ap |ψi = |ψi and Bs |ψi = |ψi. Since all the plaquette and star operators are independent (unlike in the toric code where the product of all plaquette and star stabilizer operators is the identity), one easily verifies that dim K0 = 2, independent of the height and width of the grid [19], such that one logical qubit can be stored in this space. States in K0 are topologically protected. They cannot be distinguished by any local observable and not be evolved into each other by any local unitary. We then define a Hamiltonian that imposes an energy penalty for the violation of every stabilizer condition. Let np = (1 − (σz )⊗4 )/2, np0 = (1 − (σz )⊗3 )/2, ns = (1 − (σx )⊗4 )/2, and ns0 = (1 − (σx )⊗3 )/2. The tensor products run over the qubits depicted in Figure 1. These operators have eigenvalue 0 for states that satisfy the corresponding stabilizer conditions and eigenvalue 1 for states that violate it. For some state |ψi ∈ H we say that an anyon is present at plaquette p (vertex s) if np |ψi = |ψi (ns |ψi = |ψi) and that no anyon is present if np |ψi = 0 (ns |ψi = 0), i.e. we interpret stabilizer violations as the presence of anyons. We then use the wellknown toric code Hamiltonian ! X X HKitaev = ∆ · np + ns , (1) p
s
which simply counts the total number of anyons. The code subspace K0 , which is the degenerate ground state
3 N We therefore define the logical operators X = σx and N Z= σz acting on all qubits along the paths depicted in Figure 1. Since all such products that connect the two opposite boundaries are identical up to multiplication with stabilizers, the precise form of the path does not matter. These operators commute with all the stabilizers (i.e. they are elements of the centralizer of the stabilizer group) though are not products of stabilizers. They allow thus to act non-trivially on the state stored in K0 without leaving any smoking guns in the form of anyons. They satisfy X 2 = Z 2 = I and XZ = −ZX and may therefore be seen as Pauli operators acting on the encoded qubit. At the read-out step, all the operators ns and np are measured and the presence of anyons (the syndrome) is detected. The goal is then to annihilate all anyons (either by fusing them to the vacuum or by moving them to a boundary that can absorb them) and thereby to undo the errors caused by the diffusion of the anyons. More precisely, the goal is that the total unitary formed by the natural anyon dynamics (creation, diffusion, and annihilation) plus the error correction procedure is equal to a product of stabilizers and thus acts trivially on the code space. Even more precisely, the dynamics induced by the thermal environment are in fact a probabilistic mixture of different unitary evolutions, this will be discussed in more detail later on. FIG. 1. A planar code of size L = 8. Depicted are a four-qubit plaquette operator np , a three-qubit plaquette operator np0 , a four-qubit star operator ns , and a three-qubit star operator ns0 . The logical operator X (Z) is given by any chain of Pauli operators σx (σz ) connecting the top and bottom (left and right) boundary.
of this Hamiltonian, corresponds to the anyonic vacuum. This Hamiltonian is stable against weak local perturbations in the sense that the lifting of the ground state degeneracy through such a perturbation is exponentially small in L [1]. Starting from the anyonic vacuum, if a qubit suffers a spin-flip error σx (phase-flip error σz ), two plaquette (star) anyons are created on the two adjacent plaquettes (vertices). Once an anyon exists, it can move on the surface without any further energy cost, whereby creating further spin- or phase-flip errors along its path. Two anyons of the same kind can fuse to the vacuum. On the two horizontal (vertical) boundaries of the grid in Figure 1 a single plaquette (star) anyon can be created. Similarly, the boundaries can absorb single anyons. If two anyons are jointly created from the vacuum, move around and then fuse to the vacuum, the produced error path can be expressed as a product of stabilizers and therefore acts trivially on the code subspace K0 . The same holds if a single anyon is created on a boundary and then is absorbed by the same boundary again. The only possibility to act non-trivially on the state stored in K0 is if an error path connects the two opposite boundaries.
B.
Error correction
We will study planar codes of different sizes L, by which we mean that the number of plaquette operators from top to bottom as well as the number of star operators from left to right is given by L. Consequently, there are L − 1 four-qubit plaquettes and 2 three-qubit plaquettes from left to right and thus a total of L · (L + 1) plaquettes. The total number of qubits is then given by N = 2L2 + 2L + 1. It is well-known that in the limit of large L error correction in the toric code is unambiguously possible if less than 11% of the physical qubits are subject to uncorrelated bit- and phase-flip errors [6]. However, qubit errors induced by a thermal environment are correlated due to the diffusion of the anyons. Error correction therefore typically becomes impossible if a few percent of the qubits have suffered errors [5]. The error correction step consists of three substeps. In a first step, the anyon configuration or syndrome is determined by measuring all the operators ns and np . This means that we project to the subspaces Ki with a given anyon configuration. Since there are L · (L + 1) potential locations for anyons of each kind, there are 22L·(L+1) such spaces, each of which has dimension 2 (they can be obtained from K0 by applying single-qubit errors), so H=
22L·(L+1) M −1 i=0
Ki .
(2)
4 Indeed, one verifies that 2 · 22L·(L+1) equals dim H = 2 2N = 22L +2L+1 . In a second step, a classical computation is performed on the error syndrome whose output tells how to best annihilate the anyons by fusing them with each other or moving them to a boundary of the type that can absorb them, which is then done in a third step. In a more formal language, let ρ0 denote the initial state stored in the memory with supp ρ0 ⊆ K0 . The influence of errors on this state is then captured by a quantum channel (CPTPM) Φerr . The goal of the classical computation is then, given knowledge about the error model Φerr and the syndrome (the space Ki , that is), to find a sequence of single-qubit Pauli operators Ui which corrects the errors. In other words, we want Ui to map Ki to K0 . Now let Pi denote the projector onto Ki . Formally, we can write the error-correction procedure performed on the corrupted state ρt = Φterr (ρ0 ) as P Φcorr (ρt ) = i Ui Pi ρt Pi Ui† . At the end of the day, we want the error
(3) δ(t) := ρ0 − Φcorr ◦ Φterr (ρ0 ) 1 to be as small as possible for any given encoded state ρ0 . In the corrected state (Φcorr ◦ Φterr ) (ρ0 ) no anyons are left. We therefore either have successfully corrected all errors, performed a logical X, a logical Z or both, thus Φcorr ◦ Φterr (ρ0 ) = (1 − pX )(1 − pZ ) · ρ0 + pX (1 − pZ ) · Xρ0 X + (1 − pX )pZ · Zρ0 Z + pX pZ · XZρ0 ZX (4) (assuming that the error correction procedure treats plaquette- and star-anyons independently). We therefore have δ(t) ≤ 2 · (pX (1 − pZ ) + (1 − pX )pZ + pX pZ ) ≤ 2(pX + pZ ) .
(5)
In order to obtain simple scalar functions that characterize the decay of the stored quantum information, we study the autocorrelation functions h i † X Ccorr (t) := 2−N tr X · Φcorr ◦ Φterr (X) (6) h i † Z Ccorr (t) := 2−N tr Z · Φcorr ◦ Φterr (Z) . (7) X Z The prefactor is such that Ccorr (0) = Ccorr (0) = 1, assuming that no operation is performed on the stored information if no anyons are measured. This assumption is in fact less trivial than it may seem. Performing a logical operation in the error correction step is beneficial if an odd number of logical operators have been performed by the bath. Assume that the bath induces logical errors that leave no anyons with rate r. Then, the probability that no logical error has been performed is in fact small for times t r−1 . The probability
that after time t k logical errors have been performed is given by the Poisson distribution, P (k, rt) =
(rt)k e−rt . k!
(8)
The Poisson distribution is peaked around rt, which may be an odd integer. However, the probability X k even
P (k, rt) =
1 1 + e−2rt 2
(9)
of an even number of errors is greater than 12 for any rt, such that the optimal strategy is, indeed, not to do anything if no anyon is detected. X Ccorr (t) is 1 if after error correction at time t no logical Z-operator has been applied and −1 if one has been applied (i.e. an odd number of σz -operators has been applied to any line connecting the left and right boundX ary). Therefore, Ccorr (t) = 1 − 2pZ and analogously Z Ccorr (t) = 1 − 2pX . In conclusion we have
ρ0 − Φcorr ◦ Φterr (ρ0 ) 1 X Z ≤ 1 − Ccorr (t) + 1 − Ccorr (t) . (10) We define the lifetimeτ (ε) of the memory as the maximal Z X (t) ≥ 1 − ε for all (t), Ccorr time such that min Ccorr t ≤ τ (ε), implying that δ(t) ≤ 2ε for t ≤ τ (ε). The total evolution Φcorr ◦ Φterr is a statistical mixture of different unitary evolutions. In the numerical simulations, we will in each run follow a definite unitary evoluZ (t) is at any time given by ±1. The tion, such that Ccorr probability of a certain unitary is thereby determined by the error model Φerr . Sampling over a large number of Z (t). runs, we obtain a smooth function Ccorr We say that two sequences of single-qubit Pauli operators are equivalent if they are identical up to multiplication with stabilizers. For every given anyon configuration, there are four equivalence classes of errors that produce it from the vacuum. These equivalence classes can be mapped onto each other by application of the logical operators I, X, Z, XZ. Given a syndrome, the goal is to guess the most likely equivalence class of errors that has produced it, which one allows to remove the anyons without disturbing the stored quantum information. Calculating the probabilities of the four equivalence classes is numerically too costly to be performed with current technology. We thus make the simplifying assumption that the most likely error path that has led to the given syndrome is an element of the most likely equivalence class. This may not be true for every possible anyon configuration but seems a reasonable approximation. Applying stabilizers to the most likely error path will produce further error paths with identical or slightly lower probabilities that are elements of the same equivalence class. Rather than finding the error path with maximal probability, we may equivalently find the error path with minimal weight, if we define the weight to be the negative logarithm of the probability that a certain
5
Φerr = Φ⊗N single−qubit
,
The weight of an error chain involving `x spin-flips and `z phase-flips is then 1 − pz 1 − px + `z · ln + const , px pz
1.2 1.0 0.8 0.6 0.4
Φsingle−qubit (ω) = (1 − px )(1 − pz ) · ω + px (1 − pz ) · σx ωσx + (1 − px )pz · σz ωσz + px pz · σy ωσy . (11)
`x · ln
1.4
X Ccorr
error chain has occurred. Taking the negative logarithm ensures that the weight is additive for independent error chains. This is known as the Shannon information content of an event in classical information theory [20] and up to a constant factor the only function having the additivity property. To illustrate this, let us consider a concrete simple error model. We assume that each physical qubit suffers a spinflip error with probability px and a phase-flip error with probability pz ,
(12)
allowing us to minimize (for px , pz < 12 ) `x and `x independently. The number of single-qubit Pauli operators necessary to connect two anyons with each other is given by the so-called ‘Manhattan distance’ of the anyons, i.e. the sum of the horizontal and the vertical coordinate difference of two anyons. Similarly, the weight of a chain connecting an anyon to a boundary is given by the horizontal or vertical distance. In detail, our minimal weight matching of n anyons (of one kind) then works as follows. 1. Perform a Delaunay triangulation on the set of anyon coordinates, thereby restricting the full graph of n(n−1) edges between anyons to O(n) 2 edges. Calculate the Manhattan weights of all edges in the restricted graph. 2. For every anyon, add a ‘virtual’ partner on the closer boundary able to absorb it and add an edge to the graph with weight given by the distance to the boundary.
0.2 0.0 0.07 0.075 0.08 0.085 0.09 0.095 0.10 0.105 0.11 0.115 0.12
pz X FIG. 2. The plot shows the autocorrelation function Ccorr as a function of the fraction of physical spins pz that have suffered spin-flip errors. Circles give the numerical results, the lines are guides to the eye. The curves correspond to different lattice sizes L = 32 (blue), L = 64 (magenta) and L = 128 (red). We anticipate that in the limit L → ∞ error correction is unambiguously possible for any pz < pc , with pc & 0.102.
anyons weight zero ensures that those virtual anyons that are not connected to a real one can be removed at no cost. Employing this algorithm, we obtain the decay of the Z autocorrelation function Ccorr as a function of px illustrated for different lattice sizes in Figure 2. The curves for different lattice sizes intersect for pc & 0.102, so that in the limit of L → ∞ error correction is possible for px , py < pc . Our numerically obtained value is only slightly smaller than the theoretical value pc ' 0.1094 ± 0.0002 [6]. For the uncorrected or ‘bare’ autocorrelation functions we have 2−N tr X · Φ†err (X) ≈ 0 in the whole parameter regime depicted in Figure 2, so there is a regime where our error-correction procedure is maximally beneficial, bringing the autocorrelation function from 0 to 1.
3. Connect all virtual anyons with zero-weight edges. 4. Perform a minimum-weight matching of the graph obtained this way. Points 1. and 4. are identical to the methods used in [5, 21]. We perform the Delaunay triangulation using the library Triangle [22] while for the minimal-weight perfect matching we employ the library Blossom V [23] implementing the ‘blossom’ algorithm due to Edmond’s [24]. The numerical cost of this procedure is strongly dominated by the last step. Adding the virtual anyons ensures that each real anyon can be connected to the closest edge able to absorb it and that there is always an even number of points in the graph entering the perfect matching algorithm. Giving the edges between virtual
C.
Error model
We now turn to the situation we are actually physically interested in, namely where Φterr is induced by the memory Hamiltonian H and coupling of the memory to a thermal environment. We consider a Davies weakcoupling limit [25] and briefly summarize its discussion in [5, 13, 21]. In this limit, the evolution of the memory is described by a Markovian Master equation ρ˙t = −i [H, ρt ] + L(ρt )
(13)
where the interaction between the memory and the bath is captured in the unitarity-breaking Lindblad operator
6 L. We assume that the environment is weakly coupled to the bath through single-qubit Pauli operators and thus is able to induce spin- and phase-flip errors, leading to transitions between eigenstates of H that differ only by the application of a single-qubit Pauli operator. Processes in which an energy ω is transferred from the anyonic system to the bath happen with rate γ(ω), which depends on how the bath is modeled. An expression for γ(ω) often found in the literature is given by ωn −|ω|/ω c e (14) γ(ω) = 2κn −βω 1−e and can be derived from a spin-boson model [26, 27]. In the following, we set the cutoff frequency of the bath ωc → ∞ for simplicity. A bath with n = 1 is called ‘Ohmic’, whereas one with n ≥ 2 is called ‘super-Ohmic’. Only the former case is considered in the numerical simulations [28]. To summarize, we will use βω γ(ω) = γ(0) · (15) 1 − e−βω where we think of 1/γ(0) = (2κ1 T )−1 as the relevant microscopic timescale, since the diffusion of anyons is widely determined by γ(0). Note that (14) and (15) fulfill the detailed balance condition γ(−ω) = γ(ω) · e−βω , guaranteeing that the Gibbs stateis the fixed point of the Markovian dynamics, L e−βH = 0. In such a physical model, the weight of a hypothetical error chain is not simply given by its Manhattan length, as it was the case in (12). It is in general not true that the most likely error chain is the one with the smallest number of spin flips. The number of hoppings some time ∆t after the creation of an anyon or a pair of anyons is Poisson-distributed and for large γ(0) · ∆t there is in fact a small probability that the number of spin flips is still small. Using the Manhattan distance as the weight of an error chain connecting two anyons (as done in [5, 21]) seems thus hard to justify. Rather than trying to find the most likely error chain, we therefore try to find the most likely pairing of the defects (a pairing may either be between two anyons or between an anyon and a boundary). This is not exactly equivalent to finding the most likely equivalence class of errors but should not make a relevant difference in practice and is numerically feasible. The random walk of an anyon on the grid leads to a diffusive spreading of the probability distribution. The probability of finding it at time t at a position ~r relative to its position at time t0 is 2 0 1 · e−~r /4D(t−t ) . 4πD(t − t0 )
(16)
In the case of an Ohmic bath, the diffusion constant D is basically given by the hopping rate γ(0) [5]. Similarly, the distance vector of two anyons that have been jointly created diffuses with a constant 2D since both of its ends are moving. A sensible choice for the weight of an edge
FIG. 3. The code has suffered σx errors at the black qubits in a). The syndrome measurement detects plaquette anyons A, B, and C. For the error correction procedure, we first perform a Delaunay triangulation of the full graph with edges connecting all real anyons (in the case of only three real anyons, the full graph and the triangulation coincide). Then, virtual anyons A0 , B 0 , and C 0 are added. The graph b) is then used for the perfect matching algorithm, where the obtained matching is highlighted with arrows. Error correction fails, because anyons B and C are fused and anyon A is moved to the upper boundary, thereby performing a logical X operator on the qubit stored in K0 .
between two anyons is therefore the square of their Euclidean distance, while for an edge connecting an anyon to its closer boundary we take the weight to be twice the square of the Euclidean distance. A more thorough justification of this choice can be found in Appendix A. An example for the application of the error correction algorithm is given in Figure 3.
7 III.
LONG-RANGE REPULSION BETWEEN ANYONS
and neq
1 L2
if
L2 > exp(β∆) + 1
(21)
1 L2−ε
if
ln L βA < . Lε − 1 2−ε
(22)
Putting these bounds into (20) we also find L2−ε < exp(βeeq ) + 1 < L2
(23)
for large enough L. The lifetime (18) is inverse in neq and will thus for any ε > 0 grow faster than L2−ε as L → ∞. In the case of the planar code, single anyons can be created at and absorbed by the boundaries. The anyon production rate per boundary spin is γ(−eeq ) (if the anyon density approaches the equilibrium density), while it is 2γ(−2eeq − A) for spins in the bulk. Again, eeq can be determined self-consistently from (19) and (20), where we replace L2 by L(L + 1) in (19) for the planar code. In order to find the total creation rates, these single-spin production rates have to be multiplied by 2(L + 1) and 2L2 − 1, the number of spins on the boundary and in the bulk, respectively. The total rate for production of anyons on the boundaries is then, using the bath (15), 2(L + 1) · γ(−eeq ) ' 2(L + 1)
βeeq γ(0) , exp(βeeq )
(24)
while the total rate for the production of anyons in the bulk is 4βeeq (2L2 − 1) · 2γ(−2eeq − A) ' (2L2 − 1) γ(0) . exp(2βeeq ) (25) Since the anyon density vanishes for large enough L, we neglected for these estimates that in fact only those spins that have no adjacent anyons should be considered for the anyon production rate. Using (23) with ε = 12 , we see that both the anyon production rate on the boundaries and in the bulk go to 0 as L → ∞. But which of the two will dominate for large L? Applying again (23), it is clear exp(βeeq ) of the boundary to the bulk that the ratio ' 4L rate is in fact diverging with L. The analytics are in excellent agreement with numerical simulations (Figure 4) that show an increasing ratio after a slight minimum at L ' 16. Let us note that the ratio of “active” anyons is in fact slightly different than what the above analysis suggests. The probability that an anyon created at the boundary is not reabsorbed before it gets to the second row γ(0) γ(0) is γ(0)+γ(+e ' γ(+e . Similarly, the probability that eq ) eq ) a newly created pair is not immediately re-annihilated 6γ(0) is ' γ(2e . If we are only interested in the ratio eq +A) of boundary to bulk anyons that ever get away from their place of creation, we thus obtain a further factor γ(2eeq +A) 1 6γ(+eeq ) . For the bath (15), this factor converges to 3 for large enough L. The ratio of “active” anyons created at the boundaries to “active” anyons created in the bulk exp(βe ) is thus finally 12Leq . An anyon created at the boundary has three possible fates.
8
20
ratio
15
10
5
0 100
101
102
103
104
size FIG. 4. Ratio of the number of anyons created on the boundary to the number of anyons created in the bulk if the anyon population approaches its equilibrium value. The blue circles show numerical results sampled over a time 105 · (κ1 ∆)−1 . The red curve shows the analytical prediction exp(βeeq )/4L. We have used parameters β∆ = 1/0.3 and A/∆ = 0.1 and a bath γ(ω) as in (15).
1. It can be reabsorbed by the same boundary (which will happen to most of them) or fuse with an anyon created at the same boundary, leading to a trivial operation performed on the degenerate ground space. 2. It can fuse with an anyon created at the opposite boundary or be absorbed by the opposite boundary, leading to a logical error. 3. It can fuse with an anyon created in the bulk. We will argue that the time it takes to create a logical error via possibility 2 is larger than the lifetime of the memory in the toric case and that possibility 3 has a beneficial effect on the lifetime of the memory. Combining these arguments with the fact that anyon production on the boundary surpasses the anyon production in the bulk, we conclude that the boundary has a positive net effect on the lifetime of the memory. Let us study the time it takes to create a logical error through possibility 2. There are two time-scales involved in the creation of such an error: The time it takes to create an anyon that walks during its lifetime at least once to the opposite half of the grid and the time it takes to do this walk. What is the probability that an anyon created at a boundary ever moves to the opposite half of the grid and what is the expected number of hoppings necessary for this? Up to the boundary conditions, this problem is exactly equivalent to a well-known mathematical problem called The Gambler’s Ruin. If a newly created anyon is to ever reach the opposite half, it first
has to avoid immediate reabsorbtion. The probability of γ(0) ever getting to the second row is ' γ(+e . Now imageq ) ine that the distance of the anyon to the boundary that has created it represents a gambler’s bankroll. He starts with one unit of money (the anyon starts at the first row) and then does a series of fair coin flips for one unit of money each. He ends when either he gets broke (the anyon gets reabsorbed by the same boundary that has created it) or his bankroll reaches L/2 units of money (the anyon reaches the opposite half of the grid). Since the gambler starts with one unit of money and he does a series of games with zero expectation value, his probability of ever reaching L/2 has to be 2/L, which also answers the corresponding question for the anyon. Note that movements of the anyon parallel to the boundary that has created it are irrelevant for the production of logical errors, which is why the problem can be mapped to such a one-dimensional one. Note that in the case of anyons this probability provides in fact an upper bound, since we neglected the possibility that the anyon fuses with another one before reaching the opposite half of the grid. A formally correct treatment of the problem can be found in Appendix C. There, we also show that the average number of coin flips (hoppings perpendicular to the creating boundary) needed to reach L/2 is (L/2)2 . To summarize, anyons are created at the boundaries γ(0) with rate 2(L + 1)γ(−eeq ), only a fraction γ(+e do eq ) not immediately get reabsorbed and only a fraction of at most L2 ever reaches the opposite half of the grid. If an anyon does reach the opposite grid, this takes on average 2 1 ·2· L2 (the factor 2 takes hoppings parallel to a time γ(0) the creating boundary into account). The total lifetime of a memory in which anyons are only created at the boundaries may therefore be estimated as γ(+eeq ) L 1 L2 1 · · + · 2(L + 1)γ(−eeq ) γ(0) 2 γ(0) 2 1 ' · eβeeq + 2L2 4γ(0)
(26)
where we used detailed balance. Note that the only time-scale that entered the first summand was the anyon creation time 1/γ(−eeq ). However, this cancels exactly with γ(+eeq ): the higher the anyon production rate, the higher is (by detailed balance) the probability that an anyon is immediately reabsorbed, such that the only remaining time-scale is 1/γ(0). The only property of the bath γ(ω) we used was the detailed balance property, so this scaling behavior is independent of the particulars of the bath and not specific for (15). However, in the case of a super-Ohmic bath with γ(0) = 0 the analysis of the Gambler’s Ruin problem has to be redone with an effective hopping rate emerging from indirect hopping processes that scale with γ(−2eeq − A) [5]. Since eeq diverges logarithmically with L (23), the effective hopping rate becomes vanishing for large L, leading to an improved scaling of the memory lifetime with L [5].
9
1.0
Z Ccorr (t)
0.8
0.6
0.4
0.2
0.0 100
101
102
t
units of (κ1 ∆)
103 −1
104
FIG. 5. Temporal decay of the autocorrelation functions Z Ccorr (t) for different lattice sizes and boundary conditions, for the Hamiltonian (17) and the bath (15). Times are in units of (κ1 ∆)−1 and the physical parameters are β∆ = 1/0.3 and A/∆ = 0.1. From left to right (at height 0.6, say) we have (grid type – size L) toric 32, planar 32, toric 64, planar 64, toric 128, planar 128, toric 256, planar 256, toric 512, and plaZ nar 512. The intersection of Ccorr (t) with the horizontal line at height 0.9 is used to determine the lifetimes τ (0.1) given in Figure 6.
units of (κ1 ∆)−1
104
103
τ (0.1)
Using (20) we see that the first summand in (26), i.e. the timescale needed for the creation of an anyon that walks to the opposite half, is for large enough L identical to the lifetime of the toric code (18). Furthermore, we know from (23) that the second summand grows faster than the first one. We conclude that the time needed for anyons created on the boundaries to produce a logical error grows faster with L than the time before the matching of anyons produced in the bulk becomes ambiguous. In the limit of very large L, anyons are created almost exclusively on the boundary, once the anyon production rates have approached their equilibrium values. However, as long as the anyon population is small enough anyon production in the bulk will outweigh anyon production on the boundaries, so that possibility 3 is non-negligible even for large L. Possibility 3 has the same effect as if the bulk anyon with which the boundary anyon fused had moved to and been absorbed by the creating boundary. The effect of possibility 3 is therefore to cause an effective bias of bulk anyons to move to the closer boundary. We saw that the probability that a boundary anyon gets a certain distance away from its creating boundary decreases at least inversely with that distance. This effect is thus the stronger, the closer the bulk anyons are to the boundary. We saw in the discussion of the toric case that error correction breaks down if the anyons that have been created as parts of the same pair have moved sufficiently far away from each other such that the pair-matching becomes ambiguous. If two such anyons, however distant they are, now move to the same boundary this ambiguity is resolved without a logical error. In conclusion, the equilibrium density of anyons neq is not affected by the boundary conditions, but the fraction of anyons created on the boundary becomes dominant for large enough L. The time it takes for boundary anyons to create a logical error grows faster with L than the lifetime of the toric code. We expect a beneficial effect from the possibility of boundary anyons fusing with bulk anyons. We conclude that for large enough L a planar code of size L has a larger lifetime than a toric code of size L. This is indeed confirmed by our numerical simulations. Already for L & 32 the lifetime τ (0.1) of the planar code exceeds the one of the toric code. For very small memories (L < 10) the toric code is superior, which may be attributed to logical errors caused by single anyons. In the planar code in Figure 1, for example, an anyon created at a boundary needs to perform only 4 hoppings to cause a logical error when error correction is performed. Figure 5 illustrates the temporal decay of the stored quantum information by depicting the autocorreZ lation functions Ccorr (t) for different lattice sizes and for both planar and toric grids. The obtained lifetimes are illustrated in Figure 6. While we included the analytical prediction (18) for the lifetime of the toric code in Figure 6, we cannot give such a simple expression for the lifetime of the planar
101
102
100 100
101
102
103
104
size (L) FIG. 6. Lifetimes τ (0.1) of the quantum information stored in the systems described in the caption of Figure 5. The blue crosses represent planar codes, the red circles toric codes. The sizes L of the codes are 8, 16, 32, 64, 128, 256, 512, and 1024. The red line shows the analytical prediction for the lifetime in the toric case obtained from (18) with c(0.1) = 5.1%. [5] find with an analogous plot [Fig. 6] c(0.1) = 4.4%. The increased lifetime is due the choice of the square of the Euclidean distance rather than the Manhattan length as the weight of an error chain.
10 code. When L is increased, the effect of the boundary changes from adversarial to beneficial.
IV.
A.
THE HONEYCOMB MODEL AS A QUANTUM MEMORY From the honeycomb to the planar code
The toric code Hamiltonian (1) (involving local four qubit couplings) can be realized as a low-energy effective Hamiltonian of the Kitaev honeycomb model, which involves only local two-qubit couplings [15, 16, 31]. The Hamiltonian of the honeycomb model can be simplified through a spin to hard-core boson transformation [31]. In this new language, the toric code Hamiltonian (1) emerges as a low-energy (no hard-core bosons present) fourth-order effective Hamiltonian. The qubits of the toric code obtained this way are not identical to the physical qubits of the underlying honeycomb lattice, but are effective qubits. The interactions of (1) between four of these effective qubits are mediated through processes in which two pairs of virtual hard-core bosons are created and then fuse again to the vacuum. An interaction involving three effective qubits would correspond to a process in which a pair of virtual hard-core bosons is created from the vacuum, followed by a hopping and an annihilation process. Clearly, such processes give no contribution to the effective Hamiltonian and the third-order effective Hamiltonian vanishes. There is no way of obtaining from the honeycomb model the three-qubit boundary terms introduced in [18] and used so far in this paper. Figure 7 schematically shows a planar code obtained from the honeycomb model, involving only four- but no three-qubit operators. The ground space of the Hamiltonian obtained this way, i.e. the space stabilized by all four-qubit operators (σz )⊗4 , (σx )⊗4 , has then a large degeneracy growing with the size of the memory, even if we forget about the four isolated qubits in the corners that do not interact with anything. Let G denote the group generated by all four-qubit stabilizers. In an abstract language, elements of the Pauli group on all effective qubits (except the four isolated ones in the corners) that are in the centralizer C(G) of the stabilizer group G, but not in G itself, perform logical operations on the stabilized subspace that cannot be detected through violation of a stabilizer. The goal is then to find Pauli-like observables X, Z ∈ C(G) \ G which allow us to store a qubit that is topologically protected. Elements of C(G) \ G that commute with X and Z will not do any harm to the stored qubit, but those that do not commute with X or Z can. For topological protection we therefore want • X and Z to have a macroscopic distance (i.e. to necessitate O(L) single-qubit operations), and • all elements of C(G) \ C(hX, Zi) to have macroscopic distance.
FIG. 7. An effective planar code obtained from a honeycomb lattice with a boundary. The black dots represent effective qubits emerging from the underlying honeycomb model. The four isolated effective qubits in the corners are depicted for completeness but not needed for information storage. Blue (dark) squares (like p) are plaquette operators, red (light) squares (like s) are star operators. There are two nonequivalent ways of defining a pair of macroscopic Pauli-like observables, the “big” logical operators X and Z and the “small” logical operators X 0 and Z 0 . However, an undetectable logical error consisting of only three single-qubit errors can be performed on all of these operators. For instance, a sequence of σx errors on the three qubits along path 1 (2) causes an error on the logical Z 0 (Z) operator.
Unfortunately, in the planar code obtained from the honeycomb model there is no choice of logical operators X and Z that fulfill these requirements. As illustrated in Figure 7, the pairs X, Z and X 0 , Z 0 fulfill the first requirement but not the second one. We will therefore expand G in such a way that all elements of C(G) \ G have a macroscopic distance. For the space stabilized by G being non-trivial we need G to be Abelian and to not contain −I. All of these requirements are met if we add to G the single-qubit operators σx acting on all qubits surrounded by a red circle in Figure 7 and the operators σz acting on all qubits surrounded by a blue circle. Error paths 1 and 2 in Figure 7 now no longer are elements of C(G) \ G: 1 is no longer an element of C(G) since it anti-commutes with two blue single-qubit operators in G while 2 becomes an element of G (it is the product of the two red operators enclosed by it). Therefore, the “small” operators X 0 and Z 0 now satisfy our requirements for topological protection while the “big” operators X and Z still violate the second requirement. Our new stabilized subspace topologically protects exactly one qubit. In fact, our new code is exactly equiv-
11 alent to the planar code with three-qubit operators on the boundary, since the single-qubit stabilizer operators effectively eliminate the two degrees of freedom of the qubits on the boundary. However, the Hamiltonian dynamics are different since there is no energy penalty associated with the violation of the single-qubit stabilizers on the boundary. We therefore assume that we are able to perform the measurements corresponding to the single-qubit stabilizers at the read-out step. Note that these operators act on effective qubits, but σx and σz measurements performed on them indeed correspond to measurements of a single physical qubit of the underlying honeycomb lattice. We refer to [16] for details about the mapping between physical and effective qubits. If one of the single-qubit stabilizers on the boundary is violated, we may imagine that an anyon is present at the position adjacent to it which is outside of the actual grid. For example, a σx (σz ) error can create two plaquette (star) anyons at positions p and p0 (s and s0 ) in Figure 7. An anyon at position p0 (s0 ) can only escape to p (s). In terms of anyon dynamics there are now for both kinds of anyons two boundaries that can create and absorb them and two boundaries that “attract” and store them. For example, a plaquette (star) anyon at position p (s) can reduce its energy by hopping to position p0 (s0 ). Analogously, escaping from one of these boundary positions has an energy cost. An anyon that hops to position p00 or s00 has been absorbed by the corresponding boundary and can, unlike one at position p0 or s0 , no longer be detected through a violated stabilizer operator. During error correction, an anyon stored at position p0 or s0 will be moved to the interior of the grid.
B.
Repulsion between anyons and anyon holes
As discussed, the simple toric code Hamiltonian (1) does not provide a lifetime of the stored quantum information that can be increased by making the memory larger. Pedrocchi et al. [16] studied a honeycomb model as introduced above coupled to cavity modes. The cavity modes allow the read-out of the error syndrome of the effective toric code through frequency shifts. In a resonant parameter regime, the Hamiltonian X Heff = ∆ na n ¯ a0 (27) a,a0
is Pfound perturbatively [eq. (52) in [16]]. Here, the sum a runs over all stars and plaquettes (four-qubit operators only) and n ¯ a = 1 − na counts anyon holes. For ∆ > 0 this Hamiltonian describes an effective repulsion between anyons and anyon holes. For details about how ∆ > 0 can be achieved, we refer to [16] and Section VII in [5]. The requirement to stay strictly in the perturbative regime puts an upper bound L∗ on the linear size L of the memory [16].
With Hamiltonian (27) and a total of N four-qubit operators, the gap P above the anyonic vacuum is (N − 1)∆ and as long as a na < N2 the cost to add a further anyon is at least N2 ∆. If L denotes the number of stars from top to bottom and the number of plaquettes from left to right (so L = 3 in Figure 7) we have N = 2L(L + 1). With an anyon creation gap that grows quadratically in L and a natural bath like (15) in which the creation rate decreases exponentially with the gap, we expect the production of anyon pairs in the bulk to be negligible against the creation of anyons at the boundaries. Note that if we say that a plaquette anyon is created at a boundary where single-qubit σz measurements are performed, this implies that furthermore a virtual anyon is “stored” in this boundary. From the discussion of the Gambler’s Ruin problem in Appendix C it is clear that an anyon will do at most ∼ L2 hoppings before being absorbed by or stored in a boundary. We thus expect the time an anyon is present (and not stored in a boundary) to be much smaller than the time it takes to create a single anyon. To summarize, we are already for moderate values of L in the regime −1 L2 −1 [4L · γ ((N − 1)∆)] L2 · γ (2(N − 2)∆) γ(0) (28) where almost always one or no anyon is present. In this regime, error correction works as follows. 1. Move any anyons in the interior of the grid perpendicularly to the closest of the four boundaries. (Each anyon is thereby moved to its most likely place of creation. We never assume that a pair of anyons has been created in the bulk.) 2. If an anyon is “stored” in a boundary (at position p0 or s0 in Figure 7, say), move it to the interior of the grid (to position p or s). 3. The anyons at each boundary can be matched with each other and the two-adjacent boundaries in two different not trivially suboptimal ways. Chose the one with lower weight, where the weight is given by the square of the Euclidean distance. For this system, the perfect matching problem could be reduced to a one-dimensional one. Both possible matchings that are not trivially suboptimal can be explicitly tested, such that no approximative algorithms are needed. Adding single-qubit measurements at the boundary spins ensures that in a planar code without three-body interactions all error paths that lead to a logical error after error correction consist of O(L) single-qubit errors. Figure 8 illustrates schematically all error paths of a single anyon that do lead to a logical error after error correction and all that do not. Strictly speaking, this figure is only valid for the first anyon, as a concatenation of paths that
12
1.4 1.2 1.0
Z Ccorr (n)
0.8
FIG. 8. A planar code consisting only of four- and one-qubit stabilizers. We only consider σx -errors and the corresponding plaquette operators here. As in Figure 7 σz -measurements are performed on the top and bottom boundary. The top and bottom boundary are thus able to store and detect the presence of plaquette anyons. The logical Z operator is a path of single-qubit σz operators connecting the top and bottom boundary (c.f. the operator Z 0 in Figure 7). The dashed line lies midway between the left and right boundary. All error paths in the left part of the figure will not lead to a logical error after error correction, while for the error paths in the right part of the figure a logical X operator is performed. Whether an error path of type (*) is correctable depends not only on its length, but also on its position relative to the boundaries. However, any error path of type (*) and of horizontal length below L/2 is correctable.
do not cause an error may lead to an error after error correction. It is in principal possible to find an analytical expression for the probability that the first anyon causes an error in such a planar code of size L. However, we expect from the discussion of the Gambler’s Ruin problem in Appendix C that this probability takes the form γ(0) 2 γ(0)+γ((N −1)∆) · L , up to some constant factor of order O(1). This probability will, due to both factors, be small such that, following the discussion before (10), we expect the error-corrected autocorrelation function after the creation of n single anyons to take the form Z Ccorr (n) ' 1 −
γ(0) c0 · ·n , L γ(0) + γ ((N − 1)∆)
(29)
where c0 is a constant. Of course, this approximation can Z only be valid as long as Ccorr (n) is still relatively close to 1. The numerics in Figure 9 are in excellent agreement Z with this prediction as long as Ccorr (n) & 0.5. Correspondingly, the number of single anyons that are created during the lifetime τ (ε) of the stored quantum information will be ε · L γ(0) + γ ((N − 1)∆) · . c0 γ(0)
(30)
This prediction is compared with numerical simulations in Figure 10. Since there are for each kind of anyon 2 · (2L + 1) qubits where a single anyon can be created
0.6 0.4 0.2 0.0
−0.2 100
101
102
103
104
number n of single anyons FIG. 9. A planar code of size L = 64 consisting only of four- and one-qubit stabilizers in contact with a bath with γ ((N − 1)∆) /γ(0) = 40, where N = 2L(L + 1). The vertical axis shows the number n of single anyons that have been created. The green curve shows the error-corrected autocorZ relation function Ccorr (n) and the blue curve the uncorrected autocorrelation function ((7) without Φcorr ). Both curves are sampled over 1.2 · 104 experiments. The red curve shows the analytical prediction (29) with c0 = 1.12. We see that performing a single error correction step before the read-out of the stored quantum information allows to enhance its lifetime (τ (0.1), say) by more than an order of magnitude.
104
103
102
101 100
101
102
103
size (L) FIG. 10. The number of single anyons created during the lifetime τ (ε) of the memory as a function of the lattice size L. The points show numerical results obtained for a bath with γ ((N − 1)∆) /γ(0) = 40, where N = 2L(L+1), the lines show the analytical prediction (30) with c0 = 1.12. From bottom to top we have ε = 0.1, 0.2, 0.3.
13 (including the ones on which single-qubit measurements are performed), we obtain a lifetime τ (ε) '
ε · L γ(0) + γ ((N − 1)∆) · c0 γ(0)
· [2 · (2L + 1) · γ (−(N − 1)∆)] ε exp ((N − 1)β∆) ' 0· . 4c γ(0)
−1
(31)
In the last step, we applied detailed balance and γ(0) γ ((N − 1)∆). Since N = 2L(L + 1), the lifetime of the memory grows exponentially in L2 . This provides, to the best of our knowledge, the best scaling of a quantum memory lifetime with the linear size found so far in at most three dimensions. However, we have noted that strictly speaking L is bounded through the breakdown of the perturbative treatment of the underlying honeycomb Hamiltonian. So similarly as in [13] the found scaling of the lifetime is only valid up to some optimal L∗ . Still, our optimal L∗ is generic in the sense that changing it by a small integer will not have a drastic effect on the lifetime. We note that the only time-scale that entered (31) was the time needed to create a single anyon, while γ(0) only entered through a probability. The former canceled by detailed balance such that the only remaining time-scale was the hopping time 1/γ(0). In the toric case, anyons can only be created pairwise, such that the gap above the anyonic vacuum is increased from (N − 1)∆ (as in the planar case) to 2(N − 2)∆, leading to a lifetime increasing as ∼ exp(2(N − 2)β∆), in contrast to (31). Besides complicating error correction, the realistic case of open boundaries thus also reduces the exponentially increasing factor in the lifetime of the memory to (almost) its square root. C.
Topological order at finite temperature
The above results apply to the code during its thermalization. However, it is also interesting to study its properties once it reaches thermal equilibrium. Clearly the memory will have completely decohered by this point, since the probability of the system being in any of its ground states will be equal. However, we can assess whether topological order is present. It is known that, for the non-interacting case of (1), the thermal state is not topologically ordered for any finite temperature. However, here we show that the interactions of (27) allow the topological order to remain stable for all finite temperatures at which the perturbative derivation of the Hamiltonian (27) is valid. To determine whether topological order is present we can use one of the topological order parameters designed for mixed states, such as the anyonic topological entropy [32]. This requires the plaquettes and vertices of the code to be split into three regions, A, B, and C. These can be defined arbitrarily, except that A and B must be
bounded, B must enclose A, C must enclose B and the number of plaquettes and vertices in each region must be O(L2 ). The value of the entropy depends on how well the anyon configuration within B can be used to deduce the net anyonic occupation of A. If this can be done perfectly, the entropy takes its maximum possible value (for the planar code) of 2 ln 2, signaling that the state is topologically ordered. Let us consider the thermal state of the Hamiltonian (27). The Hamiltonian is symmetric under exchange of anyons and anyon holes. Since the gap above the ground states with all anyons or all holes grows as L2 , the thermal state in the thermodynamic limit will be an equally weighted mixture of the all hole or all anyon states. As such, measurement of the occupancy of any plaquette can be used to determine the occupancies of all plaquettes. The region B used in the definition of the anyonic topological entropy can then clearly deduce the net occupancy of the region A. The value for this entropy therefore takes its maximum value of 2 ln 2, and the state is found to be topologically ordered.
V.
CONCLUSIONS
Stabilizer Hamiltonians with local interactions in 2D do not, and in 3D seem not to, offer the possibility to passively store quantum states for a time that can be made arbitrarily larger than the relevant microscopic time-scales. Inducing long-range interactions between the excitations of a 2D Hamiltonian whose ground states are topologically ordered seems thus the most promising approach towards a realistic proposal for a quantum memory. For such long-range interactions, the influence of the boundary of the memory (which every realistic memory will have) is not negligible even for large L. We discussed two Hamiltonians proposed in the recent literature and showed that for those the boundary becomes in fact dominant. Operationally, this fact becomes relevant during the error correction step before the read-out of the stored quantum state. We showed that the classical algorithm that determines how to best remove the anyonic defects has to depend on the error model, the memory Hamiltonian and the boundary conditions. With long-range repulsion between the anyons, the energy to add a further anyon increases with the number of already existing anyons, leading to a vanishing anyon production rate. If the production rates approach their equilibrium values, the production of unpaired anyons on the boundaries will outweigh the production of anyon pairs in the bulk. The influence of the boundary anyons is beneficial since they lead to an effective bias of the bulk anyons to move towards the closer boundary. We discussed how a planar code with topological protection of the stored qubit can be obtained from a honeycomb model with two-qubit Ising coupling and the ability to perform single-qubit measurements on boundary qubits. In a resonant regime of a coupling of the honey-
14 comb to cavity modes a very strong suppression of the anyon creation rate is obtained. Furthermore, most created anyons will immediately be reabsorbed by their creating boundary. In conclusion, we have found in this regime a lifetime that grows exponentially in L2 , allowing in principle to reach macroscopic storage times even at moderately high temperatures. The non-local anyon interactions in the obtained effective Hamiltonian are so strong, that the system is topologically ordered at any
finite temperature for large enough L, as long as the perturbatively derived Hamiltonian (27) describes the dynamics of the memory accurately.
[1] Alexei Kitaev. Fault-tolerant quantum computation by anyons. Annals of Physics, 303:2, 2003. [2] Zohar Nussinov and Gerardo Ortiz. A symmetry principle for topological quantum order. 2007. arXiv:condmat/0702377. [3] Robert Alicki, Mark Fannes, and Michal Horodecki. On thermalization in kitaev’s 2d model. Journal of Physics A, 42:065303, 2009. [4] Zohar Nussinov and Gerardo Ortiz. Autocorrelations and thermal fragility of anyonic loops in topologically quantum ordered systems. Phys. Rev. B, 77:064302, 2008. [5] Stefano Chesi, Beat R¨ othlisberger, and Daniel Loss. Selfcorrecting quantum memory in a thermal environment. Phys. Rev. A, 82:022305, 2010. [6] Eric Dennis, Alexei Kitaev, and John Landahl, Andrew Preskill. Topological quantum memory. Journal of Mathematical Physics, 43:4452, 2002. [7] Robert Alicki, Michal Horodecki, Pawel Horodecki, and Ryszard Horodecki. On thermal stability of topological qubit in kitaev’s 4d model. Open Syst. Inf. Dyn., 17, 2010. [8] Alastair Kay and Roger Colbeck. Quantum selfcorrecting stabilizer codes. 2008. arXiv:0810.3557v1. [9] Sergey Bravyi and Barbara Terhal. A no-go theorem for a two-dimensional self-correcting quantum memory based on stabilizer codes. New Journal of Physics, 11:043029, 2009. [10] Jeongwan Haah and John Preskill. Logical operator tradeoff for local quantum codes. 2010. arXiv:1011.3529v2. [11] Jeongwan Haah. Local stabilizer codes in three dimensions without string logical operators. Phys. Rev. A, 83:042330, 2011. [12] Sergey Bravyi and Jeongwan Haah. Energy landscape of 3d spin hamiltonians with topological order. Phys. Rev. Lett., 107:150504, 2011. [13] Sergey Bravyi and Jeongwan Haah. Analytic and numerical demonstration of quantum self-correction in the 3d cubic code. 2011. arXiv:1112.3252v1. [14] Jeongwan Haah, private communication, Jan. 2012. [15] Alexei Kitaev. Anyons in an exactly sovable model and beyond. Annals of Physics, 321:2, 2006. [16] Fabio L. Pedrocchi, Stefano Chesi, and Daniel Loss. Quantum memory coupled to cavity modes. Phys. Rev. B, 83:115415, 2011. [17] Alioscia Hamma, Claudio Castelnovo, and Claudio Chamon. Toric-boson model: Toward a topological quantum memory at finite temperature. Phys. Rev. B, 79:245122,
2009. [18] Sergey Bravyi and Alexei Kitaev. Quantum codes on a lattice with boundary. 1998. arXiv:quant-ph/9811052v1. [19] Let the ‘height’ h be given by the number of plaquette operators from top to bottom and the ‘width’ w by the number of star operators from left to right. The total number of qubits is then given by 2hw + h + w + 1, the number of plaquette operators by h(w + 1) and the number of star operators by (h + 1)w. Since each stabilizer condition eliminates half of all 22hw+h+w+1 degrees of freedom, we are left with 2 of them. [20] David J. C. MacKay. Information Theory, Inference & Learning Algorithms. Cambridge University Press, New York, NY, USA, 2002. [21] Beat R¨ othlisberger, James R. Wootton, Robert M. Heath, Jiannis K. Pachos, and Daniel Loss. Incoherent dynamics in the toric code subject to disorder. Phys. Rev. A, 85:022313, 2012. [22] J. R. Shewchuk. Applied Computational Geometry: Towards Geometric Engineering, volume 1148 of Lecture Notes in Computer Science. Springer-Verlag, 1996. [23] V. Kolmogorov. Math. Prog. Comp., 1:43, 2009. [24] J. Edmons. Can. J. Math., 17:449, 1965. [25] E. B. Davies. Comm. Math. Phys., 39:91–110, 1974. [26] A. J. Leggett, S. Chakravarty, A. T. Dorsey, M. P. A. Fisher, A. Garg, and W. Zwerger. Rev. Mod. Phys., 59:1, 1987. [27] David P. DiVincenzo and Daniel Loss. Rigorous born approximation and beyond for the spin-boson model. Phys. Rev. B, 71:035318, 2005. [28] Note that for a super-Ohmic bath we have γ(0) = 0, forbidding the direct hopping of anyons and heavily suppressing their diffusion. In this case, only ‘indirect hopping’ [5] is possible, in which a new pair of anyons is created next to an existing one and the existing one fuses with one of them, leading to an effective movement of the already existing anyon. [29] We assume in this section that the energy penalty for both three- and four-qubit stabilizer violations have the same strength ∆. This can be achieved by appropriately desigend perturbative gadgets [30]. However, we will see that the stability of the memory for large L is due to the second term in Eq. (17), i.e. the inter-anyonic interaction, and that the bare anyon energy cost ∆ becomes in fact irrelevant. In contrast, in Section IV, we will assume that the stabilizer operators are obtained from the honeycomb model such that there is no energy penalty associated with three-qubit stabilizer operators.
VI.
ACKNOWLEDGMENTS
We acknowledge useful discussions with Fabio Pedrocchi. This work was supported by the Swiss NSF, NCCR Nanoscience, and NCCR QSIT.
15 [30] Julia Kempe, Alexei Kitaev, and Oded Regev. The Complexity of the Local Hamiltonian Problem. SIAM J. Comput., 35:1070, 2006. [31] Julien Vidal, Kai Phillip Schmidt, and S´ebastien Dusuel. Perturbative approach to an exactly solved problem: Kitaev honeycomb model. Phys. Rev. B, 78:245121, 2008. [32] James R. Wootton. A witness for topological order and stable quantum memories in abelian anyonic systems. J. Phys. A: Math. Theor., 45:395301, 2012.
Appendix A: Determining the weight of the edges when anyons perform a random walk
Given an anyon at time t at position ~a = (a1 , a2 ), where a1 is the distance from the left boundary and a2 the distance from the upper boundary, what is the probability Pr [~a, t0 , t] that it has been created at a time 0 < t0 < t at the upper boundary? The answer can be obtained by summing (16) over all spins 0, . . . , L on the upper boundary, so Pr [~a, t0 , t] Z L+1 dx ' 0
2 2 0 1 · e−(a2 +(a1 −x) )/4D(t−t ) 4πD(t − t0 )
2 0 1 1 · e−a2 /4D(t−t ) = p 4 πD(t − t0 ) ( " " # #) a1 L + 1 − a1 · Erf p + Erf p . 2 D(t − t0 ) 2 D(t − t0 ) (A1)
The total probability that that an anyon that is at time t at position ~a has been created on the upper boundary is then Z t Pr [~a, t] = dt0 Pr [~a, t0 , t] · γboundary (t0 ) · ξ(t0 , t) , 0
(A2) where γboundary (t0 ) is the creation rate of anyons on the boundary at time t0 and ξ(t0 , t) is the probability that an anyon that has been created at a boundary at time t0 does still exist at time t. Unfortunately, the time integration cannot be performed in closed form even if we take γboundary and ξ to be constant and {. . .} = 2 in (A1) (corresponding to the L → ∞ limit). Still, the above analytics is enough to find reasonable weights for the edges in our error correction algorithm. First, let us note that we can find a similar expression like the one above for the situation where we have anyons at postion ~a and ~b and are interested in the probability that they have been jointly created in the bulk. becomes in this case h The exponential factor i 2 0 ~ ∼ exp −(~a − b) /8D(t − t ) . Intuitively, the distance vector of the two anyons performs a diffusive motion with diffusion constant 2D. A sensible choice for the weight of an edge between two anyons is therefore the square of their Euclidean distance, while for an edge connecting an anyon to its closer boundary we take the weight to be twice the square of the distance. All terms depending on time, the position of the anyons relative to the boundaries and the creation rates give then additive logarithmic correction terms to these weights. Note that in the matching which is obtained at the end of the algortihm the number of edges connecting two real anyons is always equal to the number of edges connecting two virtual ones. Furthermore, adding the same term to
16 the weights of all edges or multiplying all weights with the same positive term does not change the result of the algorithm. We may therefore give the edges that connect two virtual anyons a weight which takes all logarithmic correction terms into account. Formally, let d2 denote the square of the Euclidean distance of the pair we are interested in. Then, giving edges that connect two real anyons a weight α · d2 + β2 (with α > 0), edges that connect a real anyon with a virtual one a weight α · 2d2 + β1 and edges that connect two virtual anyons a weight 0 is equivalent to giving them weights d2 + (β2 − β1 )/α, 2d2 and −β1 /α respectively, which is again equivalent to giving them weights d2 , 2d2 and (β2 − 2β1 )/α respectively. Rather than calculating the term (β2 − 2β1 )/α analytically (which we cannot) we may then consider the weight of the virtual edges as a single optimization parameter of our algorithm. We have found numerically that in the systems we are interested in varying the weight of the purely virtual edges offers hardly room for improvement of the memory lifetime. As before, we will thus take their weight to be zero throughout Sec. III. However, there are certainly regimes where these weights become relevant. If, for example, the rate for creation of anyon pairs in the bulk is vanishing against the rate for creation on the boundaries, the edges between virtual anyons should be given a large weight.
is biased to move towards the boundary if it is at a position next to it. This is in particular the case immediately after the anyon has been created. We therefore study the following problem. Consider a lattice with rows 0, . . . , L − 1 and an anyon initially in row 0 (it has just been created). We are only interested in hoppings between the rows and ignore all movements within the same row. In row 0 there is a probability p˜ = γ(absorb) γ(absorb)+γ(0) that the anyon is absorbed by the adjacent boundary and a probability 1 − p˜ that it hops to row 1 (similarly in row L − 1, but this will be irrelevant here). In all other rows, the anyon hops to both adjacent rows with equal probability. What is the probability that the anyon reaches row L/2 (let L be even for simplicity) at least once before it is reabsorbed? Let pi the denote the probability to reach row L/2 from row i. The boundary conditions are p0 = (1 − p˜) · p1
(C1)
pL/2 = 1 .
(C2)
and
For 1 ≤ i ≤ L/2 − 1 we have pi = 12 (pi−1 + pi+1 ), or equivalently pi+1 − pi = pi − pi−1 . We find L/2−1
1 = pL/2 =
Appendix B: Bounds on the self-consistent anyon density
Using (19) to eliminate eeq in (20) we find 1 = neq · exp β(∆ + A(L2 neq − 1)) + 1 := f (neq ) . (B1) Since f (x) is monotonically increasing in x we have that x < neq ⇔ f (x) < 1. We have 1 1 f ( 2 ) = 2 [exp (β∆) + 1] L L 1 L2−ε
)>
1 L2−ε
· exp (βA(Lε − 1)) ,
= (L/2 − 1) · (p1 − p0 ) + p1
(C3)
Using (C1) to eliminate p1 we arrive at p0 =
1 − p˜ , (L/2 − 1) · p˜ + 1
(C4)
which for 1 − p˜ 1 (i.e. γ(0) γ(absorb)) simplifies to
(B2)
(B3)
(pi+1 − pi ) + p1
i=1
p0 ' (1 − p˜) ·
and f(
X
2 γ(0) 2 ' · . L γ(absorb) L
(C5)
Now let ni denote the expected number of hoppings (perpendicular to the boundary under interest) necessary to reach row L/2, assuming that it is eventually reached. The boundary conditions are
yielding the estimates (21) and (22).
n0 = 1 + n1
(C6)
nL/2 = 0 .
(C7)
and Appendix C: The Gambler’s Ruin
In the usual version of the Gambler’s Ruin problem, the probability of winning or losing is the same for every amount of money the gambler possesses. In the anyonic hopping problem we are interested in, the probability of moving towards the boundary is not excactly independent of the distance to the boundary. Since the rate for absorbtion of an anyon by a boundary γ(absorb) is usually much higher than the hopping rate γ(0), the anyon
For 1 ≤ i ≤ L/2 − 1 we have the recursion ni = 1 + 1 2 (ni−1 + ni+1 ) or equivalently ni+1 − ni = ni − ni−1 − 2. From this we expect a quadratic expression for ni and thus make the Ansatz ni = a · i 2 + b · i + c
(C8)
with the unique solution a = −1, b = 0 and c = (L/2)2 . We conclude that n0 = (L/2)2 .