Global Estimates of Errors in Quantum Computation by the Feynman–Vernon Formalism

The operation of a quantum computer is considered as a general quantum operation on a mixed state on many qubits followed by a measurement. The general quantum operation is further represented as a Feynman–Vernon double path integral over the histories of the qubits and of an environment, and afterward tracing out the environment. The qubit histories are taken to be paths on the two-sphere S2\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$S^2$$\end{document} as in Klauder’s coherent-state path integral of spin, and the environment is assumed to consist of harmonic oscillators initially in thermal equilibrium, and linearly coupled to to qubit operators S^z\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{S}_z$$\end{document}. The environment can then be integrated out to give a Feynman–Vernon influence action coupling the forward and backward histories of the qubits. This representation allows to derive in a simple way estimates that the total error of operation of a quantum computer without error correction scales linearly with the number of qubits and the time of operation. It also allows to discuss Kitaev’s toric code interacting with an environment in the same manner.


Introduction
Quantum computers are physical devices that manipulate quantum states to execute information-processing tasks [46,49]. To build a general-purpose quantum computer is a difficult experimental challenge where several different realizations have been proposed since the 1990ies [28,41]. While several large commercial initiatives to reach on the order of 50

Preliminaries and Relation to the Literature
The operation of an ideal quantum computer without decoherence can be cast in a language similar to information theory where the elementary operations (quantum gates) are implemented as unitary transformations on a pure quantum state, acting on a few qubits at a time [9,21]. The errors made in a quantum computation due to decoherence were first discussed quantitatively by Aharonov, Kitaev and Nisan in [3] using a formalism where the state is a density matrix and the transformations are quantum operations mapping density matrices to density matrices. This standard formalism is outlined in Sect. 3 below. In [3] the quantum operations were patterned after unitary quantum gates and assumed to factorize both over operations and between qubits which are not acted upon by the same operation. The computational power of such a quantum computer can hence be expressed as properties of networks of "noisy quantum gates". The main conclusion of [3] is that if a quantum computer of this type makes L quantum operations (operates L quantum gates in total, counted with multiplicity) each with error , then the combined error of the total operation scales as L . This estimate leads to higher accuracy needed the larger the quantum computer. It therefore points to the need for quantum error correction to make quantum computation possible [25,38,48].
Quantum or classical error correction must be formulated relative to some error model. It is for instance trivially impossible to correct the errors caused by a transmission channel which completely forgets the initial state. An important class of error models in quantum information theory, which will be discussed again below in Sects. 3 and 4, assumes that the quantum operation has a block structure where each block acts on the states of one qubit [20,59]. Such an error model is hence local both in space (physical qubits) and in time (no memory). Furthermore, in the same context it is often assumed that errors are Pauli channels which describe mis-application of operators. The analysis of such error models have been generalized to probability distribution over channel histories [12,60], which are therefore no longer local in time, but still local in space.
The originally proposed purpose of quantum computing was to simulate another quantum system of interest. Feynman argued that in general this cannot be done with a classical device [23]; experimental and theoretical progress in this direction of research was recently reviewed in [27]. Quantum supremacy is a term for similar efforts formulated in the language of computational complexity theory [32,44,53]. The objective is then to solve some classically impossible computational problem using quantum resources, or to show mathematically that an output of some quantum device needs many more resources to be simulated classically. A central model problem has here been BosonSampling [1], related to computing the permanent of a matrix. It was shown already in [61] that the permanent appears in an exact expression for the probability of scattering bosons, and that it therefore can be estimated by an ideal quantum device. In contrast, while the determinant of matrix can be computed in polynomial time e.g. by diagonalization, all general classical algorithms to compute a permanent take exponential time in the size of the matrix. Noisy BosonSampling is the task of sampling from the distribution of a number of nonideal scattered bosons (photons), whether or not that distribution is related to a permanent, or has some other application [50]. Aaronson and Arkhipov showed that Noisy Boson-Sampling remains computationally hard for somewhat abstract and small noise [1], while Kalai and Kindler showed that it becomes simple when the scattering matrix is perturbed by another small but fixed matrix [36]. Closer to the physics of the problem Aaronson and Brod showed that Noisy BosonSampling is hard when at most a finite number of the photons are dropped [2] while Oszmaniec and Brod showed that it is easy if out of n photons all but √ n are dropped [47]. It is currently unknown whether Noisy BosonSampling is hard or easy when a constant fraction (αn) of n photons are dropped.
Noisy BosonSampling mainly lies outside of the issues studied in this paper because photon drop, as a quantum problem, is on the level of second quantization. The focus in the following will be on systems composed of a fixed number of spin-1 2 fermions, or "qubits", as has been the case in most of the quantum information theory literature. As briefly reviewed in Sect. 4 such systems can also behave noisily in various ways, but for the most part the computational entities can be assumed to be long-lived, and the analysis can therefore be carried out on the level of first quantization.

Statement of the Problem
We assume that there are physical systems with Hilbert space of dimension two that we call qubits. A pure quantum state on n qubits is a complex ray in a N = 2 n -dimensional Hilbert space. A mixed state on the same n qubits is a density matrix ρ which is a non-negative Hermitian operator of unit trace; the set of all density matrices has dimension N 2 − 1. A quantum operation is a linear map from a set of density matrices to a set of density matrices which we will write Φ; this is a set of dimension N 4 − N 2 . The geometry of density matrices and general operations of a single qubit (n = 1, N = 2) are well understood as the Bloch sphere and linear transformations of the Bloch sphere, but for higher dimensions there is no such simple picture [10].
Let now n qubits start in the pure state |i >= |i 1 , . . . , i n >, density matrix ρ i = |i 1 , . . . , i n i 1 , . . . , i n |, and let there be a unitary quantum operation Φ 0 ρ i = Uρ i U † with the property that if the qubits are measured in the final state then the Boolean vector f = ( f 1 , . . . , f k ) is observed with probability P (0) i f = f |Φ 0 ρ i | f . Let then the system be coupled to an environment and described by a quantum operation Φ and corresponding probabilities P i f = f |Φρ i | f . A basic measure of the error of P i f with respect to P (0) i f is the variational distance between the two probability distributions: Any choice of final observable O taking values o leads to probability distributions P io and P io , and it can be shown the maximum of (1) over O is the trace norm · 1 of the difference of the corresponding density matrices. Furthermore, the authors of [3] introduced the diamond norm over super-operators and prove the important inequalities Φ 1 Φ 2 ♦ ≤ Φ 1 ♦ Φ 2 ♦ (Lemma 12, statement 3) and Φρ 1 ≤ Φ ♦ ρ 1 (Lemma 12, combining statements 1 and 2). For two series of quantum operations that can be written and where each pair of unitary and noisy quantum operations satisfies Φ l − Φ (0) l ♦ < this leads to [3,Theorem 4] TVD ≤ L · Equation (3) says that the total error of a quantum computer scales linearly with the number of operations L which in many realistic settings would be proportional to the number of computational units (n) and the duration of the process (t). If an error rate is defined as TVD per n and t then (3) has the interpretation that the error rate per quantum operation of a quantum computer is bounded by a constant. In particular it does not increase with the number of qubits the quantum computer is operating on. Quantum error correction systems are built on physical (small-scale) qubits which are used to build logical (larger-scale) qubits on which the quantum computation is done. Under assumptions that will be critically discussed in Sect. 4, the linear scaling (3) together with sufficiently small error-rate for individual physical qubits and gates acting on them allows quantum computing based on quantum error-correction. Kalai's "pessimistic hypothesis" [34] is the contrary position to the above. For concreteness I will formulate it as follows: Pessimistic hypothesis for a large enough quantum computational system it is not possible to maintain the scaling of the error in (3) due to correlations of errors in space and time.
If true, this would render high-quality quantum error-correction in particular, and quantum computing in general impossible for large enough systems. At this point it must be emphasized that the above statement cannot be found in [34], or in its extended version available on-line as [35], or in Kalai's earlier papers such as [33]. It is a reformulation of the pessimistic hypothesis such that the methods used in this paper are applicable. I argue in Appendices A and B that it is a reasonable reformulation. In Appendix B I also discuss briefly a model proposed [33] which can be contrasted to exact equations found by Weiss, Grifoni, Hänggi and co-workers [30].
The problem addressed in this paper is to discuss bounds on the left-hand side of (3) without considering error rates of individual components at all. The objective is hence to circumvent the critique of [34] by treating the problem as one of the physics of open quantum system, and not as one of quantum information theory. The tool to do this, used in Sects. 5-8, is the Feynman-Vernon formalism. As measurements would usually be performed in some predetermined way which would often more or less amount to measuring the z-components of all the qubits I will for simplicity assume a given initial state ρ i and a given final observable, and consider all the variability of the problem to stem from Φ being different from Φ (0) . The trace norm and the diamond norm will therefore not appear in the following analysis. Furthermore, error rate is not a concept intrinsic to quantum mechanics. In the following error rate will therefore only be discussed as an auxiliary quantity defined in terms of TVD, in the same way as done above, in text below equation (3).

Quantum Noise
As quantum noise is central to the problem addressed in this paper I will in this Section make a detour and outline the theoretical and experimental boundaries within which I discuss this concept. The reader primarily interested in the main argument may proceed directly to Sect. 5.
Quantum mechanics is based on unitary evolution of a state between measurements and non-unitary collapse of the wave function when it is measured. The latter is a source of uncertainty which is taken to be a basic property of the world [66]. Quantum computing without decoherence fully incorporates this quantum mechanical measurement uncertainty which therefore does not need to be considered further here.
Unitary time evolution shares with Hamiltonian dynamics in classical mechanics the property that it is deterministic and time-reversal invariant. It is therefore, in a colloquial sense of the word, noise-free, and does not, by itself, explain the subjective human experience that time flows forward towards the future, and not towards the past. Modifications of the equations of quantum mechanics to be stochastic were considered in [29] and more recently discussed by Weinberg [62,64]. The success of quantum mechanics as physical theory implies that such modifications, if they exist, must be very small. Such hypothetical modifications can therefore also be ignored in the present context; Weinberg in [63] gives a relative bound of 10 −17 by comparing to the stability of atomic clocks.
The issue of quantum noise is instead that a quantum mechanical system may effectively develop in a different manner than by unitary time evolution because it is interacting with another (unobserved) system. There are two ways in which this can be described: by quantum operations acting on the density matrix of the system, as summarized above in Sect. 3, or by explicitly modeling the time evolution of the observed and unobserved systems together. It is well known that every quantum operation has an environmental representation but that this is not unique; many environments and couplings to the environment correspond to the same quantum operation on the system [10].
From the point of view of information theory the simplest and most natural quantum noise models are the quantum operations that are structurally simple and most similar to unitary evolution. The factorized error model in [3] outlined in Sect. 3 assumes that the elementary unitary transformation of a noise-free quantum gate is modified to a quantum operation that acts non-trivially only on these same qubits. A more physical interpretation was given in [60] where each qubit is attached to its own separate environment (a "bath"), and then extended to the case where these baths interact when and in the same combinations as the qubits do. Other contributions have extended the model and methods of [3] to when the quantum operations depend on time [11,12,54,60]. All these contributions (and others) have in common a high level of mathematical sophistication, and the need for assumptions that are physically questionable, or at least not simple.
From the point of view of physics the simplest and most natural quantum noise models are instead those that result from simple interactions with simple environments. The simplest of these are linear interactions with an environment of harmonic oscillators. This is the model that will be introduced in Sect. 5 below and used as the basis of the subsequent analysis. Such models describe a system interacting with delocalized degrees of freedom such as photons (in cavity electrodynamics) or phonons (in solid state systems).
It is worth emphasizing that the two views on simplicity are not aligned; in fact they are more nearly orthogonal. The quantum operation that results from a harmonic oscillator bath originally in a thermal state depends strongly on bath temperature. If sufficiently high then the resulting time development of the system is Markovian i.e. factorizes over time, one of the assumptions made in [3]. In this same limit the system however behaves nearly classically [16,17], not a desired property of a quantum computing device. If on the other hand bath temperature is low then the noise from the bath acting on the system will be moderate in overall size but correlated in time and space. In the regime where a system could work as a quantum computer it must thus be able to deal with such non-trivial noise, at least as long as it may be interacting with phonons or photons.
The current leading technology for future quantum computers are coupled superconducting quantum circuits [22,65]. Each logical element (qubit) is then in fact formed by a mesoscopic object containing many millions of atoms, but where the behavior of one degree of freedom can be assimilated to that of one quantum spin. A figure-of-merit of how accurate is such a description is the ratio between the gate time of operation and the qubit relaxation time for which the current experimental (published) record for coherent super-positions is about 5 · 10 −4 . This is based on T 2 ∼ 20μs and a previously established cycle time about 10ns, alternatively one can give the number 2 · 10 −4 based on the qubit relaxation time T 1 ∼ 60μs [51]. The current (published) record for a system of nine qubits, and with all properties measured in the same system, is for one qubit (one out of nine) about 10 −3 . This number is based on measured relaxation times T 1 = 18 − 41μs and measured operation times 20 − 45ns, as given in [37, Table S3].
The qubit degree of freedom in the quantum circuit interacts with the other degrees of freedom in the circuit, with degrees of freedom in the surrounding device and material, and with as external control potential, an influence also mediated by the degrees of freedom of the device. The total dynamics is hence potentially quite complex. Deviations from desired dynamics include changes in density matrix of computational states of the qubit as well as leakage, i.e. excitations of higher non-computational states of the qubit. Considering only the first type of effects they can be modeled by interactions between a qubit and an environment, the kind of model to be introduced in Sect. 5 below. As recently reviewed in [52], fast environmental modes have to be treated quantum mechanically while slow environmental modes can be treated as classical random fields. The analysis in Sect. 5 and following hence pertain to the fast environmental modes, treated as a harmonic oscillator bath interacting linearly with the qubit.
The influence of classical random fields on the density matrix of a qubit will be a a superposition of random unitary transformations i.e. ρ → V i ρV † i , each unitary V i applied with probabilities p i . On a single qubit all such transformations can be represented as Pauli chan- are non-negative numbers that sum to one (take V i = e i π 2σ i = iσ i ). A qubit system perturbed by a Pauli channel is one of the standard models in the quantum computing literature [20,25,38,54], and the factorized error model in [3] is obviously also of the same general kind. As follows from the preceding discussion such models are not realistic descriptions of interactions with an environment: as they have no memory the corresponding environmental modes should be treated quantum mechanically. On the other hand, Pauli channels and similar models describe the effects on the quantum system of a memory-less classical uncertainty in the control potential. Note in passing that the number of independent unitary transformations in N -dimensional Hilbert space is N 2 − 1, and the dimensionality of the class of random super-positions is thus only a 1/N 2 -small fraction of all quantum operations. For instance, all random superposition of unitary transformations are unital (preserve the identity) and therefore do not include e.g. amplitude decay channels [10].
The kind of error models considered in the more recent quantum information literature which include memory [11,12,60] are more aligned with the influence of slow environmental modes. However, 1/ f -noise is an ubiquitous property of solid state devices, and this may lead stronger memory effects than have been analyzed up to now, for further discussion the interested reader is referred to [52].
Summarizing this Section, quantum noise in systems currently considered for quantum computing can be classified as (A) classical noise acting quantum mechanically, (B) influence from a slow quantum environment that can be described classically, and (C) influence from a fast quantum environment that has to be described quantum mechanically. From a fundamental point of view only (C) can be an obstacle to quantum, as opposed to classical, computing. From a practical and experimental point of view any of (A), (B) and (C) could be the main problem. Most of the quantum information literature tacitly assume (A), and as will be shown in the following analysis it is correct that (C) generally gives weaker effects than (A). The effects of (B) are more difficult to treat, likely more system dependent, and could well be main obstacles to successful quantum computing, as argued in [52].

The Open Quantum System Model
The aim of this section is to compare the two probabilities P i f and P (0) i f in one term in (1) when the quantum computer interacts with a heat bath. The Hamiltonian describing the quantum computer and the bath together iŝ whereĤ S depends only on the variables describing the quantum computer, from hereon also referred to as the system,Ĥ B depends only on the bath variables, andĤ I describes the interaction of the system and the bath. We first consider the system without the heat bath and use the observation that any unitary transformation can be implemented by unitary transformation acting on at most two qubits at a time [9,21]. The system Hamiltonian will thus beĤ where S a = {Ŝ a x ,Ŝ y a ,Ŝ z a } are the spin operators acting on the a'th qubit, μ a is a 3-vector and κ ab is a 3-by-3 matrix. Both the μ's and the κ's have dimension frequency and can depend on time as required to implement the overall unitary transformation where T means time ordering. Following the prescription of [39] we insert an over-complete resolution of the identity and write where (θ a ) parametrize unit spheres, K cs is the coherent-state propagator and |θ (i) , φ (i) and |θ ( f ) , φ ( f ) are the initial and final product coherent states. The unitary quantum operation is given by and the first matrix element we are looking for is where F means "forward" and B means "backward". The coherent-state propagator has a path integral representation where L kin and L reg are Klauder's kinetic and regularization terms, N is a normalization, and L S represents the interactions. The two coherent state propagators in (9) can therefore be written as a double path integral over forward and backward paths. I summarize for convenience the Klauder theory in Appendix C where I also give the explicit expression for L S corresponding to (5). For compactness I shall write (9) as where · · · i f is a shorthand the averages implied by (9) and (10). To the above model we now add a bath described byĤ B andĤ I . The bath and the system are originally assumed to be in a product state ρ T OT (6), depends on the whole Hamiltonian. The reduced density matrix of the system only at the final time is The total unitary operator U T OT can be represented as a path integral over the forward paths of both the system and the bath, and analogously for (U T OT ) † . The initial density matrix of the bath ρ B i can also be represented as a function of the starting points of the forward and backward paths of the bath (both to be integrated over).
As was first shown in [24] the bath variables can then be integrated out. Instead of the two coherent-state propagators in (9), each expressed as a separate path integral (10), we then instead have where is the Feynman-Vernon influence action. We can then write the second matrix element we are looking for as When the Feynman-Vernon influence action is relatively small we therefore have Extracting a representative value Φ F V we have Eq. (16) is the first result of this paper. It means that the error made by the whole system is determined by a global description of the system and the environment and is proportional to the strength of the interaction between the two. It therefore allows to estimate the scaling of the error with system size by estimating the scaling of Φ F V . In the following two section we will look at two simple models where this leads to the same scaling as Aharonov-Kitaev-Nisan (eq. 3).
and therefore use a different representation of spin histories than in [42], but the description of the bath and the coupling of the system and the bath will be the same. The model discussed here will hence be referred to as the spin-boson model with one bath per spin ("1 − 1"). For one spin the termsĤ I andĤ B in (4) are [42] H 1−1 spin-boson = nh ω n (a † n a n + 1 2 ) +Ŝ z n h 2m n ω n C n a n + c.c.
whereŜ z is the z-component of the spin and a † n and a n and the creation and annihilation operators of harmonic oscillator labeled by index n. The mass and the frequency of the harmonic oscillators are given by m n and ω n , and the strength of the interaction between the spin and the environment is given by C n . For many spins the model discussed in this section assumes one set of terms as in (17) per spin, each with a different set of operators and a † n and a n In the path integral formulation we write instead of (17) the classical Hamiltonian representing the terms involving the environment as where in the coherent-state path integral S z is the function 1 2 cos θ , as discussed in Appendix C. The Feynman-Vernon functional in (13) can then be computed explicitly as a functional of the forward and backward spin histories, as outlined in Appendix D. We will here only need the estimate of the Feynman-Vernon action for one spin coupled to one bath given in (57) and that the Feynman-Vernon actions from more than one disconnected systems add. The total Feynman-Vernon action in (13) is then estimated as where n is the number of spins, η is an overall measure of the strength of the interaction between a spin and its bath, and t f − t i is the duration of the process. Following (16) and assuming weak coupling (small η) we then have Eq. (16) is the second result of this paper. It should be read as a generalization (3) to a definite physical model where the interaction strength η is what gives rise to the elementary error , and where the number of qubits times the duration of the process (n · (t f − t i )) plays the role of of the number of "noisy operations" L. While there are similarities there are also differences. In the model used in [3] time does not enter since the system is supposed to develop unitarily between the "noisy operations". That is a somewhat unphysical assumption as any quantum system will interact with the environment to some extent, and therefore decohere continuously. On the other hand, in the model considered here the complexity of the quantum operation that implements the computational task does not enter; all else equal it does not matter how many operations are performed in the same time window as long as the form and the strength of the interaction between each qubit and its bath remains the same.

The Spin-Boson Model with One Common Bath for All Spins
A model where each spin has its own bath supposes that each spin is located in a separate material with separate delocalized degrees of freedom. Although not inconceivable one may also consider the situation where all the spins are located in the same material and interacting with the same delocalized degrees of freedom. The interaction terms between the spins and one oscillator in the common bath are then described by the action (21) where the interaction coefficients C b for simplicity have been taken the same for all spins interacting with the same bath oscillator. We can re-write the right-hand side of (21) as where n is the number of spins (qubits) and cos θ f = 1 n k cos θ f k , and same for the backward path, and then integrate out the bath oscillators. The result will be a Feynman-Vernon influence functional of the two collective coordinates of the same structure as (52) and (53), and which can be written Formally (22) scales quadratically with number of spins (qubits). However, it is physically reasonable that an increasing number of spins in the same material would take more place. One may think of either the spins are arranged along a line, or arranged on two-dimensional grid. In the first case one dimension of the system increases proportional to n while in the second case two dimensions of the system increase proportional to √ n, and in both cases the interaction coefficient C between the isolated spin and a delocalized mode can be expected to scale as 1/n. The number of modes in a small frequency interval dω will increase as n and the overall bath power spectrum J (ω) therefore decreases as 1/n. Combining these estimates one gets back the linear scaling in (20). Furthermore, the differences between the forward and backward paths are fluctuating quantities and at least in the high-temperature near-classical regime investigated in [17] one can expect cos θ f − cos θ b to scale as 1/ √ n. The real part of Φ in (22) would therefore give a contribution independent of the number of qubits while the imaginary part of Φ in (22) would give an error increasing slower than linearly.

The Toric Code in the Feynman-Vernon Theory
A canonical model of quantum computing and quantum error correction is Kitaev's toric code [38]. In the simplest version, which will be considered here, an N × M lattice of spins are located at edges in a regular lattice on the 2-torus, and operated on by operators called stabilizers In A-type stabilizers i ∈ star(i) denote the spins (edges) in the neighborhood of a vertex s and in B-type stabilizers i ∈ ∂ p denote the spins (edges) around a plaquette p; σ z i and σ x i are Pauli operators acting on spin i. All the stabilizers commute and the eigenspace of all of them measured simultaneously is four-dimensional. This Hilbert space can be identified with that of two spins, usually in this context called logical qubits, and Pauli operators on these two qubits are products of operators on the physical spins taken around the two basic circuits on the torus. Note that we are here concerned with the Kitaev code, and not the closely related quantum statistical mechanical system known as the Kitaev model. In that second case, see Eq. 29 below, the operators in Eq. 29 are terms in a Hamiltonian operator, and not measured continuously. The four-dimensional ground state of the Kitaev model is the one where the eigenvalues of A s and B p in Eq. 29 are all equal to one.
More complex versions of toric codes which can accommodate many more qubits will not be considered further here, nor the very considerable experimental challenges of actually building such systems; for a recent review, see [25]. The system under consideration hence consists of N M physical spins and a bath of harmonic oscillators developing according to (4) where in addition the stabilizers are continuously measured. The system Hamiltonian is thus where are the Pauli operators acting on the logical qubits, and C 1 , C 2 and C 1 , C 2 are the two basic cycles of the torus in respectively the vertex-centered and plaquette-centered lattice.
From these follow interaction Hamiltonians for the coherent-state representations of the histories of the logical qubits in the forward and backward paths These more complicated functions Q F and Q B play the same role for the interaction of the toric code with a bath of oscillators as the sums of the cosines in the simple model discussed in Sect. 7 above, compare Eq. (21). The influence functional is as above estimated as where Q is a typical value of Q F and Q B . A rough estimate of Q follows from assuming that each element of A in (25) is about 2 − N M 2 with a fluctuating sign, which is consistent with (26). Q(l, l ; m) in (27) The two states have the same quantum numbers (m) determined by the eigenvalues of the operators A s and B p , and the same or different quantum numbers given by the logical operators acting on the logical qubits (l and l ). It is known that the matrix elements of local operators in the ground state of the Kitaev model are exponentially small in system size [14]. For (m) = (1, 1, 1, . . . , 1) the logical qubits of the toric code are therefore almost insensitive to interactions with bath. The results of this section are positive for the Kitaev code, and it may be useful to compare other results in the literature. First, the standard view is that the Kitaev model in 2D with Hamiltonian (29) does not preserve its state when interacting with a finite-temperature heat bath [4,5,13,15,20], a result often stated as that the Kitaev model is not a stable quantum memory. That is not the same setting as considered here, as the stabilizer operators are then not continuously measured. To reproduce these results in the formalism of the present paper one which we can again estimate as 2 − N M 2 . The two terms Q F and Q B are however now sums of (2 N M ) 2 terms, and are therefore not small in system size. In this case estimate (28) hence gives essentially the same result as (20).

Discussion
In this work I have considered the error made by a quantum computer weakly coupled to an environment such that the quantum computer cannot be meaningfully described as a network of "noisy quantum gates". I have instead estimated the error by combining Klauder's path integral for spin and a Feynman-Vernon elimination of a thermal bath modeled as a set of harmonic oscillators interacting linearly with the qubits.
I have looked at three models. In the first two all qubits are computational units and all interact directly with a heat bath as in the spin-boson model [42]. In these two simpler models no error correction was considered: the goal was to see if the scaling of the overall error found by Aharonov, Kitaev and Nisan in [3] needs to be modified. The answer is negative. Instead of the error rate of a noisy quantum gate, a concept not defined for these models, the crucial parameter is the interaction strength between the system and the heat bath. If that parameter is small the total error scales at most linearly with system size (number of qubits) and time of operation -without any assumptions on locality in space and time.
The third model considered in the toric code of Kitaev [38] in 2D where additionally the physical qubits interact with a heat bath as in the spin-boson model. The computational units (logical qubits) of this model are non-localized degrees of freedom, much fewer in number than the physical qubits. The analysis brings out the fact that the states of the logical qubits are almost insensitive to interactions with a bath, at least in the ground state of the related Kitaev model where all the stabilizers (defined above) have value one. One consequence of this observation is that such an influence does not need to be corrected, as it is exponentially small in the system size. The combination of Klauder's path integral and Feynman-Vernon allows to treat together the interaction with a heat bath and other errors that can be modeled as Pauli channels, and can hence be considered an alternative to the quantum semi-group dynamics (Davies generator formalism) within which many systematic studies of this and the related Kitaev model have be performed previously [4,5,13].
A thermal bath consisting of harmonic oscillators is a model of delocalized environmental modes such as phonons. The main degrees of freedom in a real material at very low temperature, such as defects and nuclear spins, are on the other hand likely to be localized, and may be more accurately described as a spin bath [55]. For this case it may be argued that the environment of each qubit consists in a finite set of neighboring spins the effects of which would in principle also be given by a Feynman-Vernon action as in (13). Although precise estimates of this action would be more difficult to obtain, there seems to be no reason to assume that the number of environmental spins interacting with one qubit scales with the number of qubits of the quantum computer.
Finally, although the analysis is this paper has shown that fast environmental modes that have to be treated quantum mechanically are not a fundamental problem for quantum computing, there remains slow environmental modes. As long as these may be treated classically they cannot be a problem for quantum computing per se, but may nevertheless still pose very significant obstacles in practice, a point of view forcefully argued in [52]. reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

Appendix A: A Quantum-Mechanical Formulation of Kalai's
The purpose of this appendix is to argue that the total variational distance as defined above in (1), counted per qubit, is a reasonable quantum mechanical interpretation of the error rate discussed in [34]. I emphasize that this interpretation can not be found [34], but is introduced here as a way to state the problem within the theory of open quantum systems.
To do so we consider the special case where the quantum operation Φ 0 , determined by a unitary transformation U , is such that there is a single final state | f with f |Φ 0 ρ i | f = 1. Applying the quantum operation Φ 0 to |i can then be said to yield | f with certainty, and Φ 0 can then be called "noise-less". Applying the "noisy" quantum operation Φ and measuring all the qubits would on the other hand give the Boolean vector f with probability 1 − for some > 0 and a result different from f with total probability . Let now further Φ be such that the probabilities p(f ) are sensibly different from zero only when the Hamming distance betweenf and f is at most one, i.e. when at most one qubit has been flipped, and let the probability to flip any one qubit be = /n. The error rate so defined is then the same as 1 2 TVD, where TVD is defined in (1).

Appendix B: Kalai's Pessimistic Hypothesis
The purpose of this appendix is to argue that Kalai's pessimistic hypothesis claims that as introduced above in Appedix A scales linearly with number of qubits in the quantum computer. The argument proceeds by selected quotes from [34]. We start from The error rate in every realization of a universal quantum circuit scales up (at least) linearly with the number of qubits Readers of [34] will note that this statement is followed by The effort required to obtain a bounded error level for any implementation of universal quantum circuits increases (at least) exponentially with the number of qubits which is also important to Kalai's argument concerning universal quantum computers. In the present discussion, which focuses on the consequences for open quantum systems, I will however limit myself to the first part.
In [34] Kalai also argues by the example of a depolarizing one-qubit channel described by where ρ is the density matrix of a qubit (a positive Hermitian 2-by-2 matrix of unit trace), Φ is the quantum operation (a linear operator of the set of such matrices on itself) and 1 2 1 is the completely depolarized density matrix. The error rate is then taken to be p and, more generally ...error rate can be defined as the probability that a qubit is corrupted at a computation step, conditioned on it surviving up to this step which is followed by ...when we say that the rate of noise per qubit scales up linearly with the number of qubits, we mean that when we double the number of qubits in the circuit, the probability for a single qubit to be corrupted in a small time interval doubles In combination the above quotes imply that Kalai's pessimistic hypothesis states that the total error of the whole system scales at least quadratically with the number of qubits. In the interpretation used here, see Appendix A above, this is taken to mean that the total variational distance in (1) also scales at least quadratically. Similarly to the main text and Appendix A I emphasize again that this quadratic global scaling cannot be found [34] but is a consequence of the further interpretations introduced here.
The word pessimistic reflects the differences in focus between a physical and a computerscience style of analysis, alluded to from another point of view in Sect. 4 in main text. In physics one is (at least most of the time) not interested in worst-case scenarios or adversarial reasoning, but in typical behaviour of plausible models. This difference can be illustrated by considering the special case that the qubits interact among themselves only through polynomials in σ z operators, and are acted upon by terms linear in σ x (but not σ y ). As a computational device this would represent quantum annealing [8], hence somewhat restrictive, but the development of the system including interaction with a bath as in this paper can then be treated fully as the spin-boson problem [42]. The path-integral representation would involve only piece-wise constants paths. For this setting it is known that the expectation values of spin operators, equivalent for one spin to the full knowledge of the density matrix, obey exact (though complicated) generalized master equations (GME) presented e.g. in [30,Sect. 11.2.2]. The time-smoothened noise operators proposed by Kalai in e.g. Eq. 1 in [33] appear different in structure from GME. Arguments built on such hypothetical operators are therefore not persuasive in a context of definite physical models of the system-bath interaction, as is the focus of this paper.

Appendix C: The Klauder Coherent-State Path Integral for Spin
This appendix summarizes properties pertaining to the Klauder coherent-state path integral. The coherent states are defined as The two angles θ ∈ [0, π] and φ ∈ [0, 2π] parametrize the unit sphere with area 4π. The various matrix elements used in the main text and below are hence and the matrix element between two coherent states is When the two sets of angles are close this matrix element is up to terms which are small as (φ − φ) 2 and (θ − θ) 2 . Matrix elements of the operator for the z-component of spin are which when the two sets of angles are close means Similarly The vector S = 1 2 (sin θ cos φ, sin θ sin φ, cos θ ) is the radial vector of length 1 2 , polar angle θ and azimuthal angle φ. The coherent states provide a partition of the unity in an over-complete basis. Using (33) we have A time evolution operator can therefore be expressed as T e − ih Ĥ dt = n sin θ n dφ n dθ n 2π · · · |θ n+1 , φ n+1 θ n+1 , φ n+1 |e − ih t n+1 tnĤ dt |θ n , φ n θ n , φ n | · · · If it can be arranged that two consecutive sets of angles are close, the interaction term (10) for the interaction Hamiltonian (5) is, using (36) and (37), where the factorh has been included for convenience, compare (5) and (10). For the discussion below and in the main text it only matters that (40) is some definite function of the angles parametrizing the spin history.
To enforce that two consecutive sets of angles are close one uses a regularization term The matrix element (34) can then be written defining the kinetic term in (10): The regularization and interaction terms in (10) are as in (41) and (40) above. The canonical momenta conjugate to θ and φ are As discussed in [39] and in [7], when is set to zero the remaining action is first order. The function 1 2 cos θ then takes the meaning of momentum p, conjugate to φ. and the radial vector S can be written 1 2 1 − 4 p 2 cos φ, 1 2 1 − 4 p 2 sin φ, p . The Poisson brackets of the components of this vector satisfy the angular momentum relations. This suggests that in the path integral the operatorsŜ x ,Ŝ y andŜ z should be translated into these functions S x , S y and S z , which is indeed the prescription given by (36), (37), (38) and (40).
With the regularization (41) the path integral is a standard one, and could be interpreted as the mass of a particle confined to move on the surface of a sphere of fixed radius. The path integral in curved space (as is the sphere) is a well-developed topic with several complexities [57], but for the present discussion, where the "mass term" is only for regularization, one can simply interpret the integral of (41) as a time discretization and where sin 2 θ indicates e.g. mid-point prescription and the normalizing coefficient is The regularization and the normalization are a weight on the Fourier components of the spin history, and the path integral built on a discretization of (43) and (41) can therefore be written For finite this weight penalizes high Fourier components. Consider two realizations {θ k ,φ k } and {θ k ,φ k } which coincide at the two endpoints. The difference of their actions is the integral of L kin around a closed path, which in turn equals the area on the surface of the sphere circumscribed by that path [7]. This area depends only weakly on high Fourier components and the limit of zero is therefore well-behaved. as where T is the temperature of the bath. In (52) and (53) for the the forward spin history S f z = 1 2 cos θ f and for the backward S b z = 1 2 cos θ b . It is a consequence of the form of the coupling in the spin-boson model that the Feynman-Vernon action only depends on the polar angle θ and not on the azimuthal angle φ.
Assuming that J (ω) behaves as ηω s ω −s−1 c up to some large frequency and decays quickly for larger frequencies [42] k i and k r will both be proportional to η (in units of ω c ). The kernel k i will have support on a time interval of width −1 and the kernel k r will have support on a time interval of width the larger of −1 andh/k B T . Both (52) and (53) can therefore be simply estimated as where t f − t i is the duration of the process. This estimate is used above in the main text.

Appendix E: Analysis of Recovery and Quantum Error Correction
In the main text of the paper quantum error correction was not considered. The purpose of this appendix is to discuss to what extent an analysis based on Feynman-Vernon can be extended to a system with a recovery map. The general conditions for successful quantum error correction were formulated by Knill and Laflamme [40]. The starting point is a code space C of which the ground state of the Kitaev model, Sect. 8 in main text, is an example. The code space is a subset of a larger Hilbert space H called the coding space, and a super-operator Φ acts on density matrices on H. Perfect quantum error correction in C under Φ is possible if there exist another superoperator R, called a recovery operator, such that RΦ acts as the identity on all pure states | |, ∈ C. The general form of R is that of a measurement of the component in H orthogonal to C, followed by a unitary transformation. Alternatively, if Φ is represented in the Kraus form ρ → k A k ρ A † k , the condition can be formulated as conditions on dynamical operators A k acting on C [40, Theorem III.2]. If some quantum dynamics Φ on some space H admits quantum error correction therefore reduces to the question if there exists a code space C. In general this is not trivial to decide, see e.g. [45] for general rank-2 super-operators and [43] for general 2-qubit maps.
For error models often considered for the Kitaev code e.g. in [20] and [59] the above translates as follows. The coding space H is that of all the physical qubits, and Φ has a block structure where each block acts on the states of one physical qubit. Interaction with the environment has hence been assumed to lead to super-operator which is local in space (physical qubits) and also local in time (no memory), compare discussion in Sect. 4 in main text. By measuring stabilizers it is grosso modo possible to decide which unitary map was applied, and then correct for it by applying its inverse. Precision to this statement, consequences and concrete implementations have been discussed in great detail in the literature [20,25,38]. Within Feynman-Vernon theory the effects of random superposition of unitary transformations can be be described as follows: if the influence action from applying ρ → ρ = V a ρV † a is Φ a , and if this transformation is applied with probability p a , then the total influence action is 1 i log a p a e iΦ a [24]. The total variational distance from (16) is then If there are just a few unitary maps applied we are back to the same estimates as in Sect. 5, but if there are many and they contribute with random phases the resulting TVD could be smaller due to cancellations. Making additional assumptions we can also discuss the recovery map in the Kitaev code with error correction in the Feynman-Vernon formalism. First, we assume that the stabilizers are measured very often but not absolutely continuously. This is in line with proposed hardware implementations based on a system clock [25], and implies that the degrees of the freedom of both the logical qubits and the stabilizers can change between measurements. Second, the record of all the measurements of all the stabilizers is assumed known. The corresponding chain of projection operators acting on the full density matrix of the logical qubits, the stabilizers and the environment is then a coarse-grained history in the sense of Gell-Mann and Hartle [26]. Third, these coarse-grained histories are assumed to fulfill the decoherence conditions of [26]. When (if) this is so we can consider the results of the measurements as known classical time-dependent parameters and write the Feynman-Vernon path integral for the logical qubits and the environment as in Sect. 8. The difference for the spin-boson terms would then be that while at most instance of time the appropriate form is (27) (when the measured values of the stabilizers do not change), but sometimes it is (30) (when they do).