Global estimates of errors in quantum computation by the Feynman-Vernon formalism

The operation of a quantum computer is considered as a general quantum operation on a mixed state on many qubits followed by a measurement. The general quantum operation is further represented as a Feynman-Vernon double path integral over the histories of the qubits and of an environment, and afterward tracing out the environment. The qubit histories are taken to be paths on the two-sphere $S^2$ as in Klauder's coherent-state path integralof spin, and the environment is assumed to consist of harmonic oscillators initially in thermal equilibrium, and linearly coupled to to qubit operators $\hat{S}_z$. The environment can then be integrated out to give a Feynman-Vernon influence action coupling the forward and backward histories of the qubits. This representation allows to derive in a simple way estimates that the total error of operation of a quantum computer without error correction scales linearly with the number of qubits and the time of operation. It also allows to discuss Kitaev's toric code interacting with an environment in the same manner.


I. INTRODUCTION
Quantum computers are physical devices that manipulate quantum states to execute information-processing tasks [43,46]. To build a general-purpose quantum computer is a difficult experimental challenge where several different realizations have been proposed since the 1990ies [28,38]. While several large commercial initiatives to reach on the order of 50 qubits have been widely reported on recently [17,53], the current public stateof-the-art is that around ten qubits can be manipulated in the lab in a manner approaching to what would be required for a general-purpose quantum computer [34]. For more restricted computational tasks a device using about 1000 qubits has been reported to lead to important speed-up over classical algorithms [18].
Quantum computation would, if successful, upend the characterization of what is possible and not possible in classical information processing. That is its main promise, but also a main difficulty since a large quantum computer must have a substantial number of degrees of freedom, and large physical systems have a strong tendency to turn classical [30]. The underlying mechanism of this quantum-to-classical transition is the decoherence of the quantum state by interactions between the quantum computer and the rest of the world [63].
There has been a long-running polemic against the possibility of quantum computing going back at least to [6]. This paper is primarily concerned with the critique recently put forward in [32], to be discussed again in Sections II and III below. The gist of these arguments is that the standard error models considered in the quantum computing literature are too restrictive, and that quantum computing will for this reason not be possible in the real world when errors may be correlated in time and * eaurell@kth.se space. Indeed, errors in quantum computation are often discussed as mis-application of operators, which indicates a kind of classical uncertainty in the external potential controlling the quantum system, and not errors induced by coupling to another (unobserved) quantum system.
The goal of this paper is to consider this problem from a global point of view by investigating the errors caused by coupling a system of spins to a thermal bath of bosonic degrees of freedom. Such errors can be correlated over arbitrary distances at low enough temperature, but in a specific way determined by the physical interaction. The analysis is carried out by combining a coherent-state path integral representation of the dynamics of spin systems with the Feynman-Vernon method to integrate out the bath. The effect of the bath is then described by an influence functional coupling forward and backward coherent-state path integrals. This influence functional can be estimated simply when the system-bath interaction is weak.
The main result is that the errors usually considered are not an idealization but rather a worst-case scenario. It is correct that these standard error models mostly disregard correlations in time and space, but a physical heat bath, which could generate such correlations, is a comparatively simple system, and not an adversary in the sense of complexity theory. For the paradigmatic example of the Kitaev toric code it is further showed that coupling to a bath has effects exponentially small in the size of the lattice. For such a system only the errors normally considered therefore need to be corrected.
The paper is organized as follows: in Section II I position the problem in relation to the quantum computational literature and in Section III I state "pessimistic hypothesis" of Kalai in a version suitable to be analyzed by the methods used herein. In Section IV I review quantum noise, and in Section V I introduce the Feynman-Vernon double path integral as a model of quantum operations on n qubits which also interact linearly with a bath of harmonic oscillators. The system qubit histories are for-mulated using Klauder's coherent-state path integral for spin [36,55], and the Feynman-Vernon action (after integrating out the bath) are therefore interference terms between the forward and backward coherent-state paths. In Sections VI and VII I discuss two simple models where each qubit is either connected to its own environment, or where all the qubits are connected to one and the same environment. In Section VIII I treat Kitaev's toric code interacting with the same kind of environment, and in Section IX I summarize the paper. Appendices contain standard material on Klauder's path integral and the Feynman-Vernon theory, an annotated discussion of the pessimistic hypothesis as formulated in [32], and a discussion of quantum error correction and recovery, included for completeness.

II. PRELIMINARIES AND RELATION TO THE LITERATURE
The operation of an ideal quantum computer without decoherence can be cast in a language similar to information theory where the elementary operations (quantum gates) are implemented as unitary transformations on a pure quantum state, acting on a few qubits at a time [8,20]. The errors made in a quantum computation due to decoherence were first discussed quantitatively by Aharonov, Kitaev and Nisan in [3] using a formalism where the state is a density matrix and the transformations are quantum operations mapping density matrices to density matrices. This standard formalism is outlined in Section III below. In [3] the quantum operations were patterned after unitary quantum gates and assumed to factorize both over operations and between qubits which are not acted upon by the same operation. The computational power of such a quantum computer can hence be expressed as properties of networks of "noisy quantum gates". The main conclusion of [3] is that if a quantum computer of this type makes L quantum operations (operates L quantum gates in total, counted with multiplicity) each with error ǫ, then the combined error of the total operation scales as Lǫ. This estimate leads to higher accuracy needed the larger the quantum computer. It therefore points to the need for quantum error correction to make quantum computation possible [25,35,45].
Quantum or classical error correction must be formulated relative to some error model. It is for instance trivially impossible to correct the errors caused by a transmission channel which completely forgets the initial state. An important class of error models in quantum information theory, which will be discussed again below in Sections III and IV, assumes that the quantum operation has a block structure where each block acts on the states of one qubit [19,56]. Such an error model is hence local both in space (physical qubits) and in time (no memory). Furthermore, in the same context it is often assumed that errors are Pauli channels which describe mis-application of operators. The analysis of such error models have been generalized to probability distribution over channel histories [11,57], which are therefore no longer local in time, but still local in space.
The originally proposed purpose of quantum computing was to simulate another quantum system of interest. Feynman argued that in general this cannot be done with a classical device [23]; experimental and theoretical progress in this direction of research was recently reviewed in [27]. Quantum supremacy is a term for similar efforts formulated in the language of computational complexity theory [31,41,50]. The objective is then to solve some classically impossible computational problem using quantum resources, or to show mathematically that an output of some quantum device needs many more resources to be simulated classically. A central model problem has here been BosonSampling [1], related to computing the permanent of a matrix. It was shown already in [58] that the permanent appears in an exact expression for the probability of scattering bosons, and that it therefore can be estimated by an ideal quantum device. In contrast, while the determinant of matrix can be computed in polynomial time e.g. by diagonalization, all general classical algorithms to compute a permanent take exponential time in the size of the matrix. Noisy BosonSampling is the task of sampling from the distribution of a number of non-ideal scattered bosons (photons), whether or not that distribution is related to a permanent, or has some other application [47]. Aaronson and Arkhipov showed that Noisy BosonSampling remains computationally hard for somewhat abstract and small noise [1], while Kalai and Kindler showed that it becomes simple when the scattering matrix is perturbed by another small but fixed matrix [33]. Closer to the physics of the problem Aaronson and Brod showed that Noisy BosonSampling is hard when at most a finite number of the photons are dropped [2] while Oszmaniec and Brod showed that it is easy if out of n photons all but √ n are dropped [44]. It is currently unknown whether Noisy BosonSampling is hard or easy when a constant fraction (αn) of n photons are dropped.
Noisy BosonSampling mainly lies outside of the issues studied in this paper because photon drop, as a quantum problem, is on the level of second quantization. The focus in the following will be on systems composed of a fixed number of spin-1 2 fermions, or "qubits", as has been the case in most of the quantum information theory literature. As briefly reviewed in Section IV such systems can also behave noisily in various ways, but for the most part the computational entities can be assumed to be long-lived, and the analysis can therefore be carried out on the level of first quantization.

III. STATEMENT OF THE PROBLEM
We assume that there are physical systems with Hilbert space of dimension two that we call qubits. A pure quantum state on n qubits is a complex ray in a N = 2 n -dimensional Hilbert space. A mixed state on the same n qubits is a density matrix ρ which is a nonnegative Hermitian operator of unit trace; the set of all density matrices has dimension N 2 − 1. A quantum operation is a linear map from a set of density matrices to a set of density matrices which we will write Φ; this is a set of dimension N 4 − N 2 . The geometry of density matrices and general operations of a single qubit (n = 1, N = 2) are well understood as the Bloch sphere and linear transformations of the Bloch sphere, but for higher dimensions there is no such simple picture [9].
Let now n qubits start in the pure state |i >= |i 1 , . . . , i n >, density matrix ρ i = |i 1 , . . . , i n i 1 , . . . , i n |, and let there be a unitary quantum operation Φ 0 ρ i = U ρ i U † with the property that if the qubits are measured in the final state then the Boolean vector f = (f 1 , . . . , f k ) is observed with probability P (0) if = f |Φ 0 ρ i |f . Let then the system be coupled to an environment and described by a quantum operation Φ and corresponding probabili- if is the variational distance between the two probability distributions: Any choice of final observable O taking values o leads to probability distributions P io and P io , and it can be shown the maximum of (1) over O is the trace norm · 1 of the difference of the corresponding density matrices. Furthermore, the authors of [3] introduced the diamond norm over super-operators and prove the important inequalities Φ 1 Φ 2 ♦ ≤ Φ 1 ♦ Φ 2 ♦ (Lemma 12, statement 3) and Φρ 1 ≤ Φ ♦ ρ 1 (Lemma 12, combining statements 1 and 2). For two series of quantum operations that can be written and where each pair of unitary and noisy quantum operations satisfies Φ l − Φ (0) l ♦ < ǫ this leads to ( [3], Theorem 4) Equation (3) says that the total error of a quantum computer scales linearly with the number of operations L which in many realistic settings would be proportional to the number of computational units (n) and the duration of the process (t). If an error rate is defined as TVD per n and t then (3) has the interpretation that the error rate per quantum operation of a quantum computer is bounded by a constant. In particular it does not increase with the number of qubits the quantum computer is operating on. Quantum error correction systems are built on physical (small-scale) qubits which are used to build logical (larger-scale) qubits on which the quantum computation is done. Under assumptions that will be critically discussed in Section IV, the linear scaling (3) together with sufficiently small error-rate for individual physical qubits and gates acting on them allows quantum computing based on quantum error-correction. Kalai's "pessimistic hypothesis" [32] is the contrary position to the above. For concreteness I will formulate it as follows: Pessimistic hypothesis: for a large enough quantum computational system it is not possible to maintain the scaling of the error in (3) due to correlations of errors in space and time.
If true, this would render high-quality quantum errorcorrection in particular, and quantum computing in general impossible for large enough systems. At this point it must be emphasized that the above statement cannot be found in [32]; it is a reformulation of the pessimistic hypothesis such that the methods used in this paper are applicable. I argue in Appendices A and B that it is a reasonable reformulation, The problem addressed in this paper is to discuss bounds on the left-hand side of (3) without considering error rates of individual components at all. The objective is hence to circumvent the critique of [32] by treating the problem as one of the physics of open quantum system, and not as one of quantum information theory. The tool to do this, used in Sections V-VIII, is the Feynman-Vernon formalism. As measurements would usually be performed in some pre-determined way which would often more or less amount to measuring the z-components of all the qubits I will for simplicity assume a given initial state ρ i and a given final observable, and consider all the variability of the problem to stem from Φ being different from Φ (0) . The trace norm and the diamond norm will therefore not appear in the following analysis. Furthermore, error rate is not a concept intrinsic to quantum mechanics. In the following error rate will therefore only be discussed as an auxiliary quantity defined in terms of TVD, in the same way as done above, in text below equation (3).

IV. QUANTUM NOISE
As quantum noise is central to the problem addressed in this paper I will in this Section make a detour and outline the theoretical and experimental boundaries within which I discuss this concept. The reader primarily interested in the main argument may proceed directly to Section V.
Quantum mechanics is based on unitary evolution of a state between measurements and non-unitary collapse of the wave function when it is measured. The latter is a source of uncertainty which is taken to be a basic property of the world [22]. Quantum computing without decoherence fully incorporates this quantum mechanical measurement uncertainty which therefore does not need to be considered further here.
Unitary time evolution shares with Hamiltonian dynamics in classical mechanics the property that it is deterministic and time-reversal invariant. It is therefore, in a colloquial sense of the word, noise-free, and does not, by itself, explain the subjective human experience that time flows forward towards the future, and not towards the past. Modifications of the equations of quantum mechanics to be stochastic were considered in [29] and more recently discussed by Weinberg [59,61]. The success of quantum mechanics as physical theory implies that such modifications, if they exist, must be very small. Such hypothetical modifications can therefore also be ignored in the present context; Weinberg in [60] gives a relative bound of 10 −17 by comparing to the stability of atomic clocks.
The issue of quantum noise is instead that a quantum mechanical system may effectively develop in a different manner than by unitary time evolution because it is interacting with another (unobserved) system. There are two ways in which this can be described: by quantum operations acting on the density matrix of the system, as summarized above in Section III, or by explicitly modeling the time evolution of the observed and unobserved systems together. It is well known that every quantum operation has an environmental representation but that this is not unique; many environments and couplings to the environment correspond to the same quantum operation on the system [9].
From the point of view of information theory the simplest and most natural quantum noise models are the quantum operations that are structurally simple and most similar to unitary evolution. The factorized error model in [3] outlined in Section III assumes that the elementary unitary transformation of a noise-free quantum gate is modified to a quantum operation that acts non-trivially only on these same qubits. A more physical interpretation was given in [57] where each qubit is attached to its own separate environment (a "bath"), and then extended to the case where these baths interact when and in the same combinations as the qubits do. Other contributions have extended the model and methods of [3] to when the quantum operations depend on time [10,11,51,57]. All these contributions (and others) have in common a high level of mathematical sophistication, and the need for assumptions that are physically questionable, or at least not simple.
From the point of view of physics the simplest and most natural quantum noise models are instead those that result from simple interactions with simple environments. The simplest of these are linear interactions with an environment of harmonic oscillators. This is the model that will be introduced in Section V below and used as the basis of the subsequent analysis. Such models describe a system interacting with delocalized degrees of freedom such as photons (in cavity electrodynamics) or phonons (in solid state systems).
It is worth emphasizing that the two views on simplicity are not aligned; in fact they are more nearly or-thogonal. The quantum operation that results from a harmonic oscillator bath originally in a thermal state depends strongly on bath temperature. If sufficiently high then the resulting time development of the system is Markovian i.e. factorizes over time, one of the assumptions made in [3]. In this same limit the system however behaves nearly classically [15,16], not a desired property of a quantum computing device. If on the other hand bath temperature is low then the noise from the bath acting on the system will be moderate in overall size but correlated in time and space. In the regime where a system could work as a quantum computer it must thus be able to deal with such non-trivial noise, at least as long as it may be interacting with phonons or photons.
The current leading technology for future quantum computers are coupled superconducting quantum circuits [21,62]. Each logical element (qubit) is then in fact formed by a mesoscopic object containing many millions of atoms, but where the behavior of one degree of freedom can be assimilated to that of one quantum spin. A figure-of-merit of how accurate is such a description is the ratio between the gate time of operation and the qubit relaxation time for which the current experimental (published) record for coherent super-positions is about 5 · 10 −4 . This is based on T 2 ∼ 20µs and a previously established cycle time about 10ns, alternatively one can give the number 2 · 10 −4 based on the qubit relaxation time T 1 ∼ 60µs [48]. The current (published) record for a system of nine qubits, and with all properties measured in the same system, is for one qubit (one out of nine) about 10 −3 . This number is based on measured relaxation times T 1 = 18 − 41µs and measured operation times 20 − 45ns, as given in [34], Table S3.
The qubit degree of freedom in the quantum circuit interacts with the other degrees of freedom in the circuit, with degrees of freedom in the surrounding device and material, and with as external control potential, an influence also mediated by the degrees of freedom of the device. The total dynamics is hence potentially quite complex. Deviations from desired dynamics include changes in density matrix of computational states of the qubit as well as leakage, i.e. excitations of higher non-computational states of the qubit. Considering only the first type of effects they can be modeled by interactions between a qubit and an environment, the kind of model to be introduced in Section V below. As recently reviewed in [49], fast environmental modes have to be treated quantum mechanically while slow environmental modes can be treated as classical random fields. The analysis in Section V and following hence pertain to the fast environmental modes, treated as a harmonic oscillator bath interacting linearly with the qubit.
The influence of classical random fields on the density matrix of a qubit will be a a superposition of random unitary transformations i.e. ρ → V i ρV † i , each unitary V i applied with probabilities p i . On a single qubit all such transformations can be represented as Pauli channels i.e. ρ → p 0 ρ + i p iσ i ρσ i whereσ i are the Pauli matrices, and (p 0 , p 1 , p 2 , p 3 ) are non-negative numbers that sum to one (take V i = e i π 2σ i = iσ i ). A qubit system perturbed by a Pauli channel is one of the standard models in the quantum computing literature [19,25,35,51], and the factorized error model in [3] is obviously also of the same general kind. As follows from the preceding discussion such models are not realistic descriptions of interactions with an environment: as they have no memory the corresponding environmental modes should be treated quantum mechanically. On the other hand, Pauli channels and similar models describe the effects on the quantum system of a memory-less classical uncertainty in the control potential. Note in passing that the number of independent unitary transformations in N -dimensional Hilbert space is N 2 − 1, and the dimensionality of the class of random super-positions is thus only a 1/N 2 -small fraction of all quantum operations. For instance, all random superposition of unitary transformations are unital (preserve the identity) and therefore do not include e.g. amplitude decay channels [9].
The kind of error models considered in the more recent quantum information literature which include memory [10,11,57] are more aligned with the influence of slow environmental modes. However, 1/f -noise is an ubiquitous property of solid state devices, and this may lead stronger memory effects than have been analyzed up to now, for further discussion the interested reader is referred to [49].
Summarizing this Section, quantum noise in systems currently considered for quantum computing can be classified as (A) classical noise acting quantum mechanically, (B) influence from a slow quantum environment that can be described classically, and (C) influence from a fast quantum environment that has to be described quantum mechanically. From a fundamental point of view only (C) can be an obstacle to quantum, as opposed to classical, computing. From a practical and experimental point of view any of (A), (B) and (C) could be the main problem. Most of the quantum information literature tacitly as-sume (A), and as will be shown in the following analysis it is correct that (C) generally gives weaker effects than (A). The effects of (B) are more difficult to treat, likely more system dependent, and could well be main obstacles to successful quantum computing, as argued in [49].

V. THE OPEN QUANTUM SYSTEM MODEL
The aim of this section is to compare the two probabilities P if and P (0) if in one term in (1) when the quantum computer interacts with a heat bath. The Hamiltonian describing the quantum computer and the bath together isĤ whereĤ S depends only on the variables describing the quantum computer, from hereon also referred to as the system,Ĥ B depends only on the bath variables, andĤ I describes the interaction of the system and the bath. We first consider the system without the heat bath and use the observation that any unitary transformation can be implemented by unitary transformation acting on at most two qubits at a time [8,20]. The system Hamiltonian will thus bê where S a = {Ŝ a x ,Ŝ y a ,Ŝ z a } are the spin operators acting on the a'th qubit, µ a is a 3-vector and κ ab is a 3-by-3 matrix. Both the µ's and the κ's have dimension frequency and can depend on time as required to implement the overall unitary transformation where T means time ordering. Following the prescription of [36] we insert an over-complete resolution of the identity and write where (θ and the first matrix element we are looking for is where F means "forward" and B means "backward". The coherent-state propagator has a path integral representation Dθ a Dφ a e i L kin +ǫLreg+LS dt (10) where L kin and L reg are Klauder's kinetic and regularization terms, N ǫ is a normalization, and L S represents the interactions. The two coherent state propagators in (9) can therefore be written as a double path integral over forward and backward paths. I summarize for convenience the Klauder theory in App. C where I also give the explicit expression for L S corresponding to (5). For compactness I shall write (9) as where · · · if is a shorthand the averages implied by (9) and (10). To the above model we now add a bath described bŷ H B andĤ I . The bath and the system are originally assumed to be in a product state ρ T OT (6), depends on the whole Hamiltonian. The reduced density matrix of the system only at the final time is ]. The total unitary operator U T OT can be represented as a path integral over the forward paths of both the system and the bath, and analogously for (U T OT ) † . The initial density matrix of the bath ρ B i can also be represented as a function of the starting points of the forward and backward paths of the bath (both to be integrated over).
As was first shown in [24] the bath variables can then be integrated out. Instead of the two coherent-state propagators in (9), each expressed as a separate path integral (10), we then instead have where is the Feynman-Vernon influence action. We can then write the second matrix element we are looking for as When the Feynman-Vernon influence action is relatively small we therefore have Extracting a representative value Φ F V we have Eq. (16) is the first result of this paper. It means that the error made by the whole system is determined by a global description of the system and the environment and is proportional to the strength of the interaction between the two. It therefore allows to estimate the scaling of the error with system size by estimating the scaling of Φ F V . In the following two section we will look at two simple models where this leads to the same scaling as Aharonov-Kitaev-Nisan (eq. 3).

VI. THE SPIN-BOSON MODEL WITH ONE BATH PER SPIN
The first model of one spin interacting with the environment was the spin-boson model, extensively investigated in [39]. We are here concerned with general interacting spin systems and therefore use a different representation of spin histories than in [39], but the description of the bath and the coupling of the system and the bath will be the same. The model discussed here will hence be referred to as the spin-boson model with one bath per spin ("1 − 1"). For one spin the termsĤ I andĤ B in (4) are [39] H 1−1 spin-boson = n ω n (a † n a n + 1 2 ) +Ŝ z n 2m n ω n C n a n + c.c. (17) whereŜ z is the z-component of the spin and a † n and a n and the creation and annihilation operators of harmonic oscillator labeled by index n. The mass and the frequency of the harmonic oscillators are given by m n and ω n , and the strength of the interaction between the spin and the environment is given by C n . For many spins the model discussed in this section assumes one set of terms as in (17) per spin, each with a different set of operators and a † n and a n In the path integral formulation we write instead of (17) the classical Hamiltonian representing the terms involving the environment as where in the coherent-state path integral S z is the function 1 2 cos θ, as discussed in App. C. The Feynman-Vernon functional in (13) can then be computed explicitly as a functional of the forward and backward spin histories, as outlined in App. D. We will here only need the estimate of the Feynman-Vernon action for one spin coupled to one bath given in (D9) and that the Feynman-Vernon actions from more than one disconnected systems add. The total Feynman-Vernon action in (13) is then estimated as where n is the number of spins, η is an overall measure of the strength of the interaction between a spin and its bath, and t f − t i is the duration of the process. Following (16) and assuming weak coupling (small η) we then have Eq. (16) is the second result of this paper. It should be read as a generalization (3) to a definite physical model where the interaction strength η is what gives rise to the elementary error ǫ, and where the number of qubits times the duration of the process (n · (t f − t i )) plays the role of of the number of "noisy operations" L. While there are similarities there are also differences. In the model used in [3] time does not enter since the system is supposed to develop unitarily between the "noisy operations". That is a somewhat unphysical assumption as any quantum system will interact with the environment to some extent, and therefore decohere continuously. On the other hand, in the model considered here the complexity of the quantum operation that implements the computational task does not enter; all else equal it does not matter how many operations are performed in the same time window as long as the form and the strength of the interaction between each qubit and its bath remains the same.

VII. THE SPIN-BOSON MODEL WITH ONE COMMON BATH FOR ALL SPINS
A model where each spin has its own bath supposes that each spin is located in a separate material with separate delocalized degrees of freedom. Although not inconceivable one may also consider the situation where all the spins are located in the same material and interacting with the same delocalized degrees of freedom. The interaction terms between the spins and one oscillator in the common bath are then described by the action where the interaction coefficients C b for simplicity have been taken the same for all spins interacting with the same bath oscillator. We can re-write the right-hand side of (21) as n 2 t f ti dtC n x n (t)cos θ f (t) where n is the number of spins (qubits) and cos θ f = 1 n k cos θ f k , and same for the backward path, and then integrate out the bath oscillators. The result will be a Feynman-Vernon influence functional of the two collective coordinates of the same structure as (D4) and (D5), and which can be written (22) Formally (22) scales quadratically with number of spins (qubits). However, it is physically reasonable that an increasing number of spins in the same material would take more place. One may think of either the spins are arranged along a line, or arranged on two-dimensional grid. In the first case one dimension of the system increases proportional to n while in the second case two dimensions of the system increase proportional to √ n, and in both cases the interaction coefficient C between the isolated spin and a delocalized mode can be expected to scale as 1/n. The number of modes in a small frequency interval dω will increase as n and the overall bath power spectrum J(ω) therefore decreases as 1/n. Combining these estimates one gets back the linear scaling in (20). Furthermore, the differences between the forward and backward paths are fluctuating quantities and at least in the high-temperature near-classical regime investigated in [16] one can expect cos θ f − cos θ b to scale as 1/ √ n. The real part of Φ in (22) would therefore give a contribution independent of the number of qubits while the imaginary part of Φ in (22) would give an error increasing slower than linearly.

VIII. THE TORIC CODE IN THE FEYNMAN-VERNON THEORY
A canonical model of quantum computing and quantum error correction is Kitaev's toric code [35]. In the simplest version, which will be considered here, an N ×M lattice of spins are located at edges in a regular lattice on the 2-torus, and operated on by operators called stabilizers In A-type stabilizers i ∈ star(i) denote the spins (edges) in the neighborhood of a vertex s and in B-type stabilizers i ∈ ∂p denote the spins (edges) around a plaquette p; σ z i and σ x i are Pauli operators acting on spin i. All the stabilizers commute and the eigenspace of all of them measured simultaneously is four-dimensional. This Hilbert space can be identified with that of two spins, usually in this context called logical qubits, and Pauli operators on these two qubits are products of operators on the physical spins taken around the two basic circuits on the torus. Note that we are here concerned with the Kitaev code, and not the closely related quantum statistical mechanical system known as the Kitaev model. In that second case, see Eq. 29 below, the operators in Eq. 29 are terms in a Hamiltonian operator, and not measured continuously. The four-dimensional ground state of the Kitaev model is the one where the eigenvalues of A s and B p in Eq. 29 are all equal to one.
More complex versions of toric codes which can accommodate many more qubits will not be considered further here, nor the very considerable experimental challenges of actually building such systems; for a recent review, see [25]. The system under consideration hence consists of N M physical spins and a bath of harmonic oscillators developing according to (4) where in addition the stabilizers are continuously measured. The system Hamiltonian is thus where are the Pauli operators acting on the logical qubits, and C 1 , C 2 and C ′ 1 , C ′ 2 are the two basic cycles of the torus in respectively the vertex-centered and plaquette-centered lattice.
A basis of the states of the physical spins is |i 1 , i 2 , . . . , i N M where i p = ±1 denotes the up (down) state of spin p. An alternative basis is by above given by the k values of the z-components of logical qubits l r (here k = 2) and the N M −k values of the stabilizers m q = ±1. These two bases are related by a unitary transformation where |l, m > denotes |l 1 , . . . , l k , m 1 , . . . , m N M−k > and |i > denotes |i 1 , i 2 , . . . , i N M >. By orthogonality of the states of the stabilizers and the logical qubits we have From these follow interaction Hamiltonians for the coherent-state representations of the histories of the logical qubits in the forward and backward paths These more complicated functions Q F and Q B play the same role for the interaction of the toric code with a bath of oscillators as the sums of the cosines in the simple model discussed in Section VII above, compare Eq. (21). The influence functional is as above estimated as where Q is a typical value of Q F and Q B . A rough estimate of Q follows from assuming that each element of A in (25) is about 2 − N M 2 with a fluctuating sign, which is consistent with (26). Q(l, l ′ ; m) in (27) is then the sum of 2 N M terms of fluctuating signs, each of size about 2 −N M and hence of overall typical size 2 − N M 2 . Each of the two functions Q F and Q B is then a sum of (2 k ) 2 such terms multiplied by the matrix elements with the angles which each have RMS average 2 − k 2 (see Appendix C). The approximate sizes of Q F and Q B are hence 2 (28) is consequently 2 k−N M , and the influence of the bath on the states of the logical qubits exponentially small in system size. A more systematic estimate of Q follows from observing that Q(l, l ′ ; m) is the matrix element of the operators coupling the system to the heat bath between two eigenstates the Kitaev model given by Hamiltonian The two states have the same quantum numbers (m) determined by the eigenvalues of the operators A s and B p , and the same or different quantum numbers given by the logical operators acting on the logical qubits (l and l ′ ). It is known that the matrix elements of local operators in the ground state of the Kitaev model are exponentially small in system size [13]. For (m) = (1, 1, 1, . . . , 1) the logical qubits of the toric code are therefore almost insensitive to interactions with bath. The results of this section are positive for the Kitaev code, and it may be useful to compare other results in the literature. First, the standard view is that the Kitaev model in 2D with Hamiltonian (29) does not preserve its state when interacting with a finite-temperature heat bath [4,5,12,14,19], a result often stated as that the Kitaev model is not a stable quantum memory. That is not the same setting as considered here, as the stabilizer operators are then not continuously measured. To reproduce these results in the formalism of the present paper one should promote the measured values of the stabilizers m to be quantum variables represented in a larger coherent-state path integral by forwards and backwards angles {θ m,F φ m,F } and {θ m,B φ m,B }. Instead of (27) we then have (m and m' different) which we can again estimate as 2 − N M 2 . The two terms Q F and Q B are however now sums of (2 N M ) 2 terms, and are therefore not small in system size. In this case estimate (28) hence gives essentially the same result as (20).

IX. DISCUSSION
In this work I have considered the error made by a quantum computer weakly coupled to an environment such that the quantum computer cannot be meaningfully described as a network of "noisy quantum gates". I have instead estimated the error by combining Klauder's path integral for spin and a Feynman-Vernon elimination of a thermal bath modeled as a set of harmonic oscillators interacting linearly with the qubits.
I have looked at three models. In the first two all qubits are computational units and all interact directly with a heat bath as in the spin-boson model [39]. In these two simpler models no error correction was considered: the goal was to see if the scaling of the overall error found by Aharonov, Kitaev and Nisan in [3] needs to be modified. The answer is negative. Instead of the error rate of a noisy quantum gate, a concept not defined for these models, the crucial parameter is the interaction strength between the system and the heat bath. If that parameter is small the total error scales at most linearly with system size (number of qubits) and time of operationwithout any assumptions on locality in space and time.
The third model considered in the toric code of Kitaev [35] in 2D where additionally the physical qubits interact with a heat bath as in the spin-boson model. The computational units (logical qubits) of this model are non-localized degrees of freedom, much fewer in number than the physical qubits. The analysis brings out the fact that the states of the logical qubits are almost insensitive to interactions with a bath, at least in the ground state of the related Kitaev model where all the stabilizers (defined above) have value one. One consequence of this observation is that such an influence does not need to be corrected, as it is exponentially small in the system size. The combination of Klauder's path integral and Feynman-Vernon allows to treat together the interaction with a heat bath and other errors that can be modeled as Pauli channels, and can hence be considered an alternative to the quantum semi-group dynamics (Davies generator formalism) within which many systematic studies of this and the related Kitaev model have be performed previously [4,5,12].
A thermal bath consisting of harmonic oscillators is a model of delocalized environmental modes such as phonons. The main degrees of freedom in a real material at very low temperature, such as defects and nuclear spins, are on the other hand likely to be localized, and may be more accurately described as a spin bath [52]. For this case it may be argued that the environment of each qubit consists in a finite set of neighboring spins the effects of which would in principle also be given by a Feynman-Vernon action as in (13). Although precise estimates of this action would be more difficult to obtain, there seems to be no reason to assume that the number of environmental spins interacting with one qubit scales with the number of qubits of the quantum computer.
Finally, although the analysis is this paper has shown that fast environmental modes that have to be treated quantum mechanically are not a fundamental problem for quantum computing, there remains slow environmental modes. As long as these may be treated classically they cannot be a problem for quantum computing per se, but may nevertheless still pose very significant obstacles in practice, a point of view forcefully argued in [49]. The purpose of this appendix is to argue that the total variational distance as defined above in (1), counted per qubit, is a reasonable quantum mechanical interpretation of the error rate ǫ discussed in [32]. I emphasize that this interpretation can not be found [32], but is introduced here as a way to state the problem within the theory of open quantum systems.
To do so we consider the special case where the quantum operation Φ 0 , determined by a unitary transformation U , is such that there is a single final state |f with f |Φ 0 ρ i |f = 1. Applying the quantum operation Φ 0 to |i can then be said to yield |f with certainty, and Φ 0 can then be called "noise-less". Applying the "noisy" quantum operation Φ and measuring all the qubits would on the other hand give the Boolean vector f with probability 1 − ǫ ′ for some ǫ ′ > 0 and a result different from f with total probability ǫ ′ . Let now further Φ be such that the probabilities p(f ) are sensibly different from zero only when the Hamming distance betweenf and f is at most one, i.e. when at most one qubit has been flipped, and let the probability to flip any one qubit be ǫ = ǫ ′ /n. The error rate so defined is then the same as 1 2 TVD, where TVD is defined in (1).

Appendix B: Kalai's pessimistic hypothesis
The purpose of this appendix is to argue that Kalai's pessimistic hypothesis claims that ǫ as introduced above in App. A scales linearly with number of qubits in the quantum computer. The argument proceeds by selected quotes from [32]. We start from "The error rate in every realization of a universal quantum circuit scales up (at least) linearly with the number of qubits' Readers of [32] will note that this statement is followed by "The effort required to obtain a bounded error level for any implementation of universal quantum circuits increases (at least) exponentially with the number of qubits" which is also important to Kalai's argument concerning universal quantum computers. In the present discussion, which focuses on the consequences for open quantum systems, I will however limit myself to the first part.
In [32] Kalai also argues by the example of a depolarizing one-qubit channel described by where ρ is the density matrix of a qubit (a positive Hermitian 2-by-2 matrix of unit trace), Φ is the quantum operation (a linear operator of the set of such matrices on itself) and 1 2 1 is the completely depolarized density matrix. The error rate is then taken to be p and, more generally "...error rate can be defined as the probability that a qubit is corrupted at a computation step, conditioned on it surviving up to this step" which is followed by "...when we say that the rate of noise per qubit scales up linearly with the number of qubits, we mean that when we double the number of qubits in the circuit, the probability for a single qubit to be corrupted in a small time interval doubles" In combination the above quotes imply that Kalai's pessimistic hypothesis states that the total error of the whole system scales at least quadratically with the number of qubits. In the interpretation used here, see App. A above, this is taken to mean that the total variational distance in (1) also scales at least quadratically. Similarly too the main text and App. A I emphasize again that this quadratic global scaling cannot be found [32] but is a consequence of the further interpretations introduced here. The two angles θ ∈ [0, π] and φ ∈ [0, 2π] parametrize the unit sphere with area 4π. The various matrix elements used in the main text and below are hence and the matrix element between two coherent states is When the two sets of angles are close this matrix element is up to terms which are small as (φ ′ − φ) 2 and (θ ′ − θ) 2 . Matrix elements of the operator for the z-component of spin are which when the two sets of angles are close means The vector S = 1 2 (sin θ cos φ, sin θ sin φ, cos θ) is the radial vector of length 1 2 , polar angle θ and azimuthal angle φ. The coherent states provide a partition of the unity in an over-complete basis. Using (C2) we have ↑ | ↑ = sin θdφdθ 2π ↑ |θ, φ θ, φ| ↑ = 1 ↑ | ↓ = sin θdφdθ 2π ↑ |θ, φ θ, φ| ↓ = 0 ↓ | ↑ = sin θdφdθ 2π ↓ |θ, φ θ, φ| ↑ = 0 ↓ | ↓ = sin θdφdθ 2π ↓ |θ, φ θ, φ| ↓ = 1 A time evolution operator can therefore be expressed as T e − i Ĥ dt = n sin θ n dφ n dθ n 2π · · · |θ n+1 , φ n+1 θ n+1 , φ n+1 |e − i t n+1 tnĤ dt |θ n , φ n θ n , φ n | · · · (C7) If it can be arranged that two consecutive sets of angles are close, the interaction term (10) for the interaction Hamiltonian (5) is, using (C4) and (C5), L S = − 1 2 a µ z a cos θ a + (µ x a cos φ a + µ y a sin φ a ) − 1 4 ab κ zz ab cos θ a cos θ b + cos θ a sin θ b (κ zx ab cos φ b +κ zy ab sin φ b ) + (κ xz ab cos φ a + κ yz ab sin φ a ) sin θ a cos θ b + sin θ a sin θ b (κ xx ab cos φ a cos φ b κ xy ab cos φ a sin φ b + κ yx ab sin φ a cos φ b + κ yx ab sin φ a cos φ b + +µ y a sin φ a ) (µ where the factor has been included for convenience, compare (5) and (10). For the discussion below and in the main text it only matters that (C8) is some definite function of the angles parametrizing the spin history.
To enforce that two consecutive sets of angles are close one uses a regularization term L reg = 1 2 θ 2 + sin 2 θφ 2 ) (C9) The matrix element (C2) can then be written defining the kinetic term in (10): The regularization and interaction terms in (10) are as in (C9) and (C8) above. The canonical momenta conjugate to θ and φ are p θ = ǫθ p = p φ = ǫ sin 2 θφ + 1 2 cos θ (C12) As discussed in [36] and in [7], when ǫ is set to zero the remaining action is first order. The function 1 2 cos θ then takes the meaning of momentum p, conjugate to φ. and the radial vector S can be written 1 2 1 − 4p 2 cos φ, 1 2 1 − 4p 2 sin φ, p . The Poisson brackets of the components of this vector satisfy the angular momentum relations. This suggests that in the path integral the operatorsŜ x ,Ŝ y andŜ z should be translated into these functions S x , S y and S z , which is indeed the prescription given by (C4), (C5), (C6) and (C8).
With the regularization (C9) the path integral is a standard one, and ǫ could be interpreted as the mass of a particle confined to move on the surface of a sphere of fixed radius. The path integral in curved space (as is the sphere) is a well-developed topic with several complexities [54], but for the present discussion, where the "mass term" is only for regularization, one can simply interpret the integral of (C9) as a time discretization and where sin 2 θ indicates e.g. mid-point prescription and the normalizing coefficient is The regularization and the normalization are a weight on the Fourier components of the spin history, and the path integral built on a discretization of (C11) and (C9) can therefore be written For finite ǫ this weight penalizes high Fourier components. Consider two realizations {θ k ,φ k } and {θ ′ k ,φ ′ k } which coincide at the two endpoints. The difference of their actions is the integral of L kin around a closed path, which in turn equals the area on the surface of the sphere circumscribed by that path [7]. This area depends only weakly on high Fourier components and the limit of zero ǫ is therefore well-behaved.