Single-shot decoding of good quantum LDPC codes

Quantum Tanner codes constitute a family of quantum low-density parity-check (LDPC) codes with good parameters, i.e., constant encoding rate and relative distance. In this article, we prove that quantum Tanner codes also facilitate single-shot quantum error correction (QEC) of adversarial noise, where one measurement round (consisting of constant-weight parity checks) suffices to perform reliable QEC even in the presence of measurement errors. We establish this result for both the sequential and parallel decoding algorithms introduced by Leverrier and Z\'emor. Furthermore, we show that in order to suppress errors over multiple repeated rounds of QEC, it suffices to run the parallel decoding algorithm for constant time in each round. Combined with good code parameters, the resulting constant-time overhead of QEC and robustness to (possibly time-correlated) adversarial noise make quantum Tanner codes alluring from the perspective of quantum fault-tolerant protocols.


INTRODUCTION
Quantum error correcting (QEC) codes [1,2] are the backbone of quantum fault-tolerant protocols needed to reliably operate scalable quantum computers.Due to their simplicity, stabilizer codes [3], which can be realized by measuring a set of commuting Pauli operators known as parity checks, have received much attention.From the perspective of fault tolerance, it might be desirable to further require that qubits are placed on some lattice and to restrict parity checks to be constantweight and geometrically local.However, such topological QEC codes, which include the toric code [4,5] and the color code [6][7][8] as examples, have limited code parameters [9][10][11].To avoid these limitations, one can drop the assumption about geometric locality of parity checks (while still maintaining the assumption about their constant weight) to obtain a more general family of QEC codes known as quantum low-density parity-check (QLDPC) codes; see Ref. [12] for a recent review.Importantly, QLDPC codes can have essentially optimal parameters, as shown by recent breakthrough results [13][14][15][16], culminating in the construction of (asymptotically) good QLDPC codes whose encoding rates and relative distances are constant [17].A key component of the construction of asymptotically good QLDPC codes is the presence of "product-expanding" local codes.Since then, a few alternative constructions of good QLDPC codes have been proposed [18,19].Good parameters alone are not enough for QEC codes to be interesting beyond the theoretical realm.In order to be practically relevant and useful, QEC codes need computationally efficient decoding algorithms which process the error syndrome and identify errors afflicting the encoded information.Importantly, decoding algorithms need to operate at least at the speed at which quantum faulttolerant protocols are being implemented; otherwise, the error syndrome will keep accumulating and one will suffer from the so-called backlog problem [20].Recently, a few computationally efficient (and provably correct) decoding algorithms have been developed for good QLDPC codes [19,21,22], assuming access to the noiseless error syndrome.
To extract the error syndrome, one usually implements appropriate quantum circuits composed of basic quantum operations, such as state preparation, entangling gates and measurements.Unfortunately, these basic operations are imperfect and, for that reason, the assumption about the noiseless error syndrome is unrealistic.In particular, practical QEC codes and decoding algorithms should exhibit robustness to measurement errors.Arguably, one of the simplest ways to achieve such robustness involves repeating measurements until a reliable account of the error syndrome is obtained [5,23].However, this approach incurs significant time overhead since the number of repetitions needed in general grows with the code distance.
An alternative to repeated measurement rounds of the error syndrome was introduced in the form of single-shot QEC by Bombín [24].The basic idea behind single-shot QEC is to carefully select a code for which the decoding problem has sufficient structure to reliably infer qubit errors even with imperfect syndrome measurements.The strength of this approach is that significantly fewer measurements are necessary for codes that admit single-shot decoding compared to the simple strategy of repeated measurements.
Single-shot QEC can be considered either for stochastic or adversarial noise.In the stochastic case, one is interested in noise that afflicts a (randomly selected) constant fraction of qubits.Additional structure may be needed for both the noise and the code, since the expected weight of the errors can be far beyond the code distance.Examples of such structure include sufficiently high expansion in the associated factor graphs, e.g., quantum expander codes [25]; or the presence of geometrically local redundancies among constant-weight parity checks, e.g., the 3D subsystem toric code [26,27] and the gauge color code [28].In the adversarial case, as considered by Campbell [29], one can realize single-shot QEC for any code by measuring a carefully chosen set of parity checks; similar ideas of exploiting a redundant set of parity checks to simultaneously handle measurement and qubit errors were also explored in Refs.[30][31][32].The limitation of this approach is that, even when starting with a QLDPC code, the parity checks needed for single-shot QEC may have weight growing with code length, which makes it less appealing from the perspective of quantum faulttolerant protocols.
We remark that while stochastic noise and adversarial noise models are generally incomparable, the distinction fades for asymptotically good QEC codes.Since these codes, by definition, have constant relative distance, they have the ability to correct arbitrary errors of weight up to a constant fraction of the number of qubits.In particular, stochastic noise with sufficiently low rate is correctable with high probability.Since in the rest of the paper we focus on good QLDPC codes, it suffices to consider the case of adversarial noise.

Main Results
In this article, we focus on a class of asymptotically good QLDPC codes called quantum Tanner codes [18].They admit computationally efficient decoding algorithms, such as the sequential and parallel mismatch decomposition algorithms introduced in Ref. [33] and the potential-based decoder introduced in Ref. [22].The problem of decoding quantum Tanner codes has so far been considered only in the scenario with noiseless error syndrome.Here, we study the performance of the aforementioned sequential and parallel mismatch decomposition decoders in the presence of measurement errors.We show that the decoders are single-shot, under the following definition.For a more detailed discussion of single-shot decoding, see Section 3. Suppose a data error e occurs on the qubits.Let σ be the (ideal) syndrome corresponding to the data error.Suppose that the measured syndrome is corrupted by measurement error D. With access to the noisy syndrome σ = σ + D as input, the decoder tries to output a correction f close to the data error.Definition 1.1 (Informal Statement of Definition 3.3).A decoder is said to be (α, β)-single-shot if, for sufficiently low-weight errors, the correction f returned on input σ satisfies |e + f | R ≤ α|e| R + β|D|, where |e| R is the stabilizer-reduced weight of e, i.e., the weight of the smallest error equivalent to e up to the addition of stabilizers.
In other words, using a single round of noisy syndrome measurement, the decoder finds and applies the correction f , resulting in the residual error e + f of weight below α|e| R + β|D|.Let n be the number of physical qubits of the quantum Tanner code.Our main theorems are as follows.
We further consider the situation where multiple rounds of qubit error, noisy syndrome measurement, and decoding occur.We show that under mild assumptions on the weights of qubit and measurement errors, repeated applications of an (α, β)-single-shot decoder will keep the residual error weight bounded.Specifically, consider the case where an initial error (e 1 , D 1 ) is partially corrected by the decoder, leaving a residual error e ′ 1 .A new error (e 2 , D 2 ) is then applied on top of the existing residual error, giving total error (e ′ 1 + e 2 , D 2 ).The decoder attempts to correct using a new round of syndrome measurements (without using the syndromes of previous rounds), leaving residue e ′ 2 .This process is repeated for multiple rounds.Then we have the following.A direct implication of this result is that for the parallel decoder (Algorithm 3), a constant number of iterations suffices to keep the residual error weight bounded at each round.This process can be repeated essentially indefinitely until ideal error correction is required, at which point the O(log n)-iteration parallel decoder can be used.For more details, see the discussion at the end of Section 3.3.
The rest of this paper is organized as follows.In Section 2, we provide the necessary background on quantum Tanner codes.For more detailed explanations, see Refs.[18] and [33].In Section 3, we describe the decoding problem for quantum (CSS) codes under measurement noise, and discuss the notion of single-shot decoding.We then define (α, β)-single-shot decoding and derive general consequences of this definition under multiple rounds of error and decoding.The main result of this section is the proof of Theorem 3.5.Section 4 forms the bulk of the paper.There, we review the sequential and parallel decoders from Ref. [33] and prove that the decoders are single-shot in Theorems 4.17 and 4.20.Finally, we end with some discussions in Section 5.

Classical codes
A classical binary linear code is a subspace C ⊆ F n 2 .We refer to n as the block length of the code.The number of encoded bits (also referred to as the code dimension) is given by k = dim C and the rate of the code is R = k/n.The distance of C is defined as d = min x∈C\{0} |x|, where | • | is the Hamming weight of a vector and where 0 denotes the zero vector.A code with distance d can protect against any unknown error of weight less than d/2.Often, it is useful to specify a code C via a parity check matrix H.By definition, C = ker H.
The dual code of a code C is defined as , where the codewords can be thought of as matrices such that every column is a codeword of C A and every row is a codeword of C B .The dual tensor code of C A and C B , denoted by C A ⊞ C B , is defined as A parity check matrix for The dual tensor codes we use are required to satisfy the following robustness condition.
Here, ∥c∥ A denotes the number of non-zero columns in c and ∥r∥ B denotes the number of non-zero rows in r.When it is clear from context, we will drop the subscripts on the norms.The notion of product-expansion was introduced by Panteleev and Kalachev [17].It is equivalent to robust testability of tensor product codes [34] and agreement testability [35], and also implies another notion called w-robustness of dual tensor codes [18].It has been proven that random codes are product-expanding with high probability [19,36].

Quantum codes
An n-qubit quantum code is a subspace C of an n-qubit Hilbert space, i.e., C ⊆ C 2 ⊗n .We are interested in stabilizer codes, which are codes that can be expressed as the simultaneous +1eigenspace of an abelian subgroup S of the n-qubit Pauli group satisfying −I ̸ ∈ S. If S can be generated by two sets S X and S Z comprising, respectively, Pauli X-type and Z-type operators, then we refer to the corresponding stabilizer code as a Calderbank-Shor-Steane (CSS) code [37,38].By ignoring the phase factors for such X-type and Z-type operators, we can identify them with their supports as vectors in F n 2 .
For any CSS code stabilized by S = ⟨S X , S Z ⟩, we can define two n-bit classical codes C X = ker H X and C Z = ker H Z , where each row in H X and H Z is the support of a stabilizer generator in S X and S Z , respectively.The dimension of a CSS code is We now describe the construction of quantum Tanner codes.The code is placed on a geometric object called the left-right Cayley complex.Let G be a finite group and A = A −1 , B = B −1 be two symmetric generating sets of G.The left-right Cayley complex Cay 2 (A, G, B) is a two-dimensional object with vertices V , edges E, and faces Q defined as follows: • Q = {{(g, 00), (ag, 01), (gb, 10), (agb, 11)} : g ∈ G, a ∈ A, b ∈ B} .
Let Q(v) denote the set of faces incident to a given vertex v.Each face incident to v can be obtained by choosing an A-type edge and a B-type edge incident to v and completing them into a square.Therefore, Q(v) is in bijection with the set A × B, and can be thought of as a matrix with rows indexed by A and columns indexed by B (Figure 1).Similarly, the set of faces incident to a given A-edge is in bijection with B and the set of faces incident to a given B-edge is in bijection with A.
Consider the usual Cayley graph Cay(A, G) with the vertex set G and the edge set {{g, ag} : g ∈ G, a ∈ A}.Ignoring the B edges from the complex, we have that (V, E A ) is the disjoint union of two copies of the bipartite cover of Cay(A, G).Similarly, (V, E B ) is the disjoint union of two copies of the bipartite cover of Cay(G, B). 1 We say that a ∆-regular graph is Ramanujan if the second largest eigenvalue of its adjacency matrix is at most 2 √ ∆ − 1, and we will consider left-right Cayley complexes with component Cayley graphs Cay(A, G) and Cay(G, B) that are Ramanujan.Explicitly, Ramanujan Cayley graphs can be obtained by taking G = PSL 2 (q i ), where q is an odd prime power and A, B are (appropriately chosen) symmetric generating sets of constant size ∆ = |A| = |B| = q + 1 [35].Quantum Tanner codes are CSS codes defined by placing qubits on the faces of a left-right Cayley complex.We fix two classical codes, C A of length |A| and C B of length |B|, which are used to define a pair of local codes providing the parity checks of the quantum code.An X-type stabilizer generator is defined as a codeword from a generating set of C 0 = C A ⊗ C B , with support on the faces incident to a given vertex in V 0 = V 00 ∪ V 11 .More precisely, there is an X-type stabilizer generator s(x, v) for every generator x ∈ C A ⊗ C B and every vertex v ∈ V 0 .Identifying Q(v) with A × B using the bijection explained earlier, the support of s(x, v) is the subset of Q(v) defined by the support of x; see Figure 2 for an illustration.Similarly, the Z-type stabilizers are generated by codewords of The fact that X and Z parity checks commute is because X and Z generators are either disjoint or overlap on the faces incident to a single edge.On this set of faces, isomorphic to either B or A, the supports of the X and Z operators are codewords of either C B and C ⊥ B , respectively, or C A and C ⊥ A , respectively.It is clear that a family of quantum Tanner codes is QLDPC if the degrees of the component Cayley graphs are bounded.
Leverrier and Zémor showed that quantum Tanner codes defined on expanding left-right Cayley complexes using product-expanding local codes have good parameters [18,33].
Theorem 2.3 (Theorem 1 in Ref. [33]).Let ρ, d r , κ ∈ (0, 1) and ∆ be a sufficiently large constant.Let C A , C B ⊆ F ∆ 2 be classical codes of rates ρ and (1 − ρ) respectively, such that the distances of C A , C B , C ⊥ A , C ⊥ B are all at least d r ∆, and such that C A ⊞ C B and C ⊥ A ⊞ C ⊥ B are both κ-productexpanding.Using a family of ∆-regular Ramanujan Cayley graphs Cay(A, G) and Cay(G, B), define the left-right Cayley complex Cay 2 (A, G, B).Then the quantum Tanner codes defined using the components above have parameters n .

Decoding CSS codes
Let us now formally define the decoding problem for quantum (CSS) codes.After we encode logical information in a quantum code, errors will occur on the physical system.We are interested in how to "undo" these errors and, subsequently, recover the original logical state.Specifically, consider a logical state |ψ⟩ of a stabilizer code C. A Pauli error E occurs, and we gain information about the error by measuring a set of stabilizer generators {S i }.This gives a syndrome σ, a bit string whose values σ i correspond to the eigenvalues (−1) σ i of the stabilizers measured.Thus, σ i = 0 whenever S i commutes with E and σ i = 1 when it anticommutes.The task of decoding is to use σ to determine a correction F such that F E|ψ⟩ = |ψ⟩.In other words, F E should be a stabilizer of the code.When C is a CSS code, we can express the problem as follows.
Definition 3.1.Let C be a CSS code specified by two parity check matrices H X ∈ F r X ×n 2 and H Z ∈ F r Z ×n

2
. Let e = (e X , e Z ) ∈ F 2n 2 be an error with corresponding syndrome σ = (σ X , σ Z ) ∈ F r Z +r X

2
, where σ Z = H X e Z and σ X = H Z e X .Given input σ, the task of decoding is to find corrections f = ( fX , fZ ) ∈ F 2n 2 such that e X + fX ∈ C ⊥ X and e Z + fZ ∈ C ⊥ Z .
In the definition above, we associate the bit string e = (e X , e Z ) with the Pauli errors E = E X E Z where E X and E Z are Pauli X and Z operators with support e X and e Z , respectively (ignoring phase information).The correction f is similarly associated with a Pauli operator F .
We note that for CSS codes, the decoding problem can be split into two separate problems for the X and Z codes that can be solved independently.For quantum Tanner codes in particular, there is symmetry between the X and Z codes, as can be seen by switching V 0 and V 1 labels and switching Therefore, it suffices to give an algorithm for decoding one type of error.In the remainder of the paper, we will consider solely the case where X-errors occur, with Z-errors treated analogously.For convenience, we will often drop subscripts, for example writing e for e X or H for H Z .
The above discussion assumes that the ideal syndrome is accessible to the decoder.Let us now consider the case when the syndrome measurements are unreliable, motivated by the fact that the quantum circuits implementing the parity checks are necessarily imperfect.Suppose that the ideal syndrome σ X of an error e X is corrupted by measurement error D X , so that the actual noisy syndrome readout is σX = σ X + D X .A naive decoding of the syndrome σX may result in a correction fX which does not bring the state back to the code space, i.e., e X + fX ̸ ∈ C ⊥ X .Furthermore, there may be no guarantee that e X + fX is close to C ⊥ X .
One of the standard procedures to account for measurement errors is to repeatedly measure the stabilizer generators in order to gain enough confidence in their measurement outcomes [5,23].This will incur large time overhead.Alternatively, syndrome measurements can be performed fault-tolerantly by preparing special ancilla qubit states offline [39,40].This will incur large qubit overhead.It would be ideal if we could avoid both overheads at the same time.

Single-shot decoding
Bombín [24] introduced single-shot decoders as an alternative approach.These decoders take in a noisy syndrome as input and, even in the presence of syndrome noise, return a correction that can be used to reduce the data error.Most likely, there will be some resulting residual error, but its weight is bounded by some function of the syndrome noise.In more detail, the single-shot property posits that it suffices to perform O(n) parity check measurements (in the context of QLDPC codes, one further requires constant weight of measured parity checks), and, using only these measurement outcomes, one can perform reliable QEC that keeps the residual noise at bay.
In our analysis, we need the following definition.The stabilizer-reduced weight of an error is a convenient theoretical measure of how detrimental the error really is.Note that since stabilizers do not change the code state, errors are only welldefined up to the addition of stabilizers.As such, any bound on the performance of the decoder is unambiguously defined using the stabilizer-reduced weight, which can be significantly smaller than the original weight.
Since we focus on asymptotically good QLDPC codes, it is enough to consider single-shot decoding for adversarial noise.Campbell [29] captures adversarial single-shot decoding as follows.Let both the data error e and the syndrome noise D be sufficiently small.A decoder is single-shot if it outputs a correction such that the weight of the residual error is bounded by a polynomial of |D|.In this work, we would like to consider constant-time decoding using the parallel decoder (Algorithm 3) for quantum Tanner codes.This setting does not directly fit into the previous definition since the residual error could depend on |e| in addition to |D|.To allow for nontrivial dependence on |e|, we give the following definition, which is relevant for asymptotically good codes where the residual error size is at most linear in |e| and |D|.
be a syndrome error, and σ = (σ X , σZ ) ∈ F r Z +r X 2 be the corresponding noisy syndrome, where σX = H Z e X + D X and σZ = H X e Z + D Z .A decoder for C is (α, β)-single-shot if there exist constants A, B, C such that, for P ∈ {X, Z}, whenever This definition, combined with Theorems 4.17 and 4.20 below, gives the following results for the sequential and parallel decoders of the quantum Tanner codes.

The parallel decoder (
Note that the runtime of the sequential decoder is O(n), and each iteration of the parallel decoder is constant time.For the parallel decoder, α decreases exponentially with the number of parallel decoding iterations k, and the results of this section will hold when k is a sufficiently large constant.It suffices to take k = O(log n) for α = 0 in the parallel decoder.
Finally, we remark that we may increase the robustness to measurement errors and improve the overall performance of single-shot decoding by leveraging redundancies among parity checks, similar to the ideas explored in Refs.[30][31][32].We can apply this approach to quantum Tanner codes without compromising their QLDPC structure, which is a crucial difference between our setting and the aforementioned works.Specifically, stabilizer generators of quantum Tanner codes are supported on local neighborhoods, defined by the local codes C 0 and C 1 .We may apply the technique of adding redundancy to each set of local checks separately.Since the local codes are of length ∆ 2 , any redundant check in a fixed local neighborhood will not have weight more than ∆ 2 , which is comparable to the weight of the original checks.

Multiple rounds of decoding
In this section, we discuss what happens after multiple rounds of errors, noisy measurements, and decoding.We show that under the assumptions of Definition 3.3, there exists a variety of noise models such that, as long as the overall noise level is sufficiently small, the encoded quantum information will persist for an exponential number of rounds.
The results proven in this section hold for any decoder that can solve the single-shot decoding problem under Definition 3.3.More precisely, we assume that if the decoder is given the noisy syndrome from data error e ∈ F n 2 and syndrome error then it outputs a correction f such that the residual error satisfies We will assume that β is constant and that α is a parameter in the decoder that can be made arbitrarily small.For our analysis, we let R, S be constants such that We prove that as long as the data and syndrome errors in each round are sufficiently small, the total error can be kept small indefinitely.
Theorem 3.5.Consider errors (e i , D i ) that occur on rounds i = 1, 2, • • • , with decoding in between each round using new syndrome measurements (i.e., without using the previous syndromes).If the errors satisfy |e i | ≤ Rn and |D i | ≤ Sn for every round i, then the residual error e ′ i after each round i satisfies Proof.Initially, e ′ 0 = 0, which satisfies the bound.Suppose after round i − 1, the residual error e ′ i−1 satisfies (2).The new total error is e ′ i−1 + e i , and we have where the last inequality follows since for R and S satisfying (1).Therefore, the decoder returns a correction f with residual error where the third inequality uses the inductive hypothesis.
From this result, we can immediately analyze the stochastic setting in which large errors are unlikely.
Corollary 3.6.Let {(e i , D i )} M i=1 be randomly distributed data and syndrome errors (with possible correlations) such that for constants a, b > 0. Suppose the decoder is run after each round of errors using new syndrome measurements (i.e., without using the syndromes of previous rounds).Then the final residual error e ′ M satisfies Proof.This follows immediately from Theorem 3.5 after using a union bound on the probability of a large data or syndrome error at every round.
As a sample application of Corollary 3.6, we analyze the case of p-bounded noise [25,41], although any model of errors with sufficiently suppressed tails will give the same conclusions.
Definition 3.7 (p-bounded noise).Let p ∈ [0, 1).Let A be a set and let 2 A be its power set.We say that a probability distribution E : 2 A → [0, 1] is p-bounded if for any B ⊆ A we have Corollary 3.8.Let {(e i , D i )} M i=1 be data and syndrome errors where each of the marginal distributions of e i and D i are p-and q-bounded, respectively.Suppose the decoder is run after each round of errors using a new round of syndrome measurements (without using the syndromes of previous rounds).Then, the final residual error e ′ M satisfies where where the last inequality follows by p-boundedness.Using the binary entropy bound for the binomial coefficient, we then have Similarly, we have Applying Corollary 3.6 gives the result.
In particular, there exist thresholds (p * , q * ) = (2 −H(R)/R , 2 −ϱH(S/ϱ)/S ) below which errors are kept under control for an exponential number of rounds of single-shot QEC with high probability.
Finally, we comment on the last round of QEC.In a typical setting of fault tolerance, we choose to measure logical qubits in the computational basis, which for a CSS code can be accomplished by measuring each physical qubit (also in the computational basis).We then apply one final round of QEC, where the Z-stabilizer eigenvalues are inferred by multiplying the Z-measurement outcomes from those qubits in the stabilizer supports.Note that in this final round, any measurement error can be treated as an X data error immediately before the measurement.We run the decoder with α sufficiently small so that by the guarantee on the decoder, |e + f | R = 0, i.e., we completely correct the error.We can then infer the logical information by combining the corrected single-qubit Z-measurement outcomes making up the Z-logical operators.Therefore, fault tolerance may be achieved by using a faster (e.g., constant-time) decoder with larger α value in the middle of the computation, and only applying the full decoder (e.g., logarithmic-time) with α = 0 at the end of the computation.
Note that this is a completely local operation which can be done without consideration of the syndrome state of the other vertices.Each square q ∈ Q contains two V 1 vertices, say v ∈ V 01 and v ′ ∈ V 10 .These two vertices are each associated with their own local corrections, εv and εv ′ , which may disagree on whether there is an error on q.If there is no disagreement on any square q ∈ Q, then a global correction f ∈ F Q 2 can be unambiguously defined by f = However, this will usually not be the case.The disagreement between the different candidate local corrections is captured by a "noisy mismatch vector" defined as The goal of the main part of the algorithm is to reduce the size of Z by successively updating the best local corrections on the V 1 vertices.For example, it is possible that for a given v ∈ V 1 , replacing εv with εv + x for some x ∈ C ⊥ 1 in (3) would significantly decrease | Z|.In general, we attempt to decompose Z by adding codewords x ∈ C ⊥ 1 on local views Q(v) of vertices v ∈ V . 2 We keep track of the decomposition process through quantities Ĉ0 , Ĉ1 , R0 , R1 ⊆ F Q 2 , which are initially 0 and updated as follows.Suppose Then we add c to Ĉj and r to Ri .The interpretation is that Ĉ1 + R0 is the total change made to the local corrections εv from the V 01 vertices, and Ĉ0 + R1 is the total change made to those from the V 10 vertices.Therefore, at the end of the procedure, we output a guess for the error, which from the perspective of the V 01 vertices is f = The algorithm can run either sequentially (Algorithm 1) or in parallel (Algorithm 3), with the corresponding Z decomposition subroutines presented in Algorithm 2 and Algorithm 4 respectively.

Algorithm 1 Sequential decoder for quantum Tanner codes with parameter ε
Input: A noisy syndrome σ arising from data error e and syndrome error D. Output: A correction f that approximates e.
1: εv ← argmin{|y| : Algorithm 2 Sequential mismatch decomposition with parameter ε for (i, j) ∈ {0, 1} 2 do 4: if there exists 0 Choose x v such that |x v | maximal among all possible choices 7: Find end parallel for each 13: end for 14: end repeat 15: return ( Ĉ0 , Ĉ1 , R0 , R1 ) These algorithms were analyzed in the scenario with perfect measurement outcomes in Ref. [33], giving the following results: Theorem 4.1 (Theorem 13 in Ref. [33]).Let ε ∈ (0, 1).Suppose Algorithm 1 with parameter ε is given as input the noiseless syndrome σ = H Z e of an error e ∈ F Q 2 of weight Then it will output a correction f such that e + f ∈ C ⊥ X in time O(n).
Theorem 4.2 (Theorem 20 in Ref. [33]).Let ε ∈ (0, 1/6).Suppose Algorithm 3 is given as input the noiseless syndrome σ = H Z e of an error e ∈ F Q 2 of weight Then it will output a correction f such that e + f ∈ C ⊥ X in time O(log n).
In the next sections, we will consider what happens when the decoders are given a syndrome with possible errors.

Proof Preliminaries
We first give a summary of the main ideas of the proof.The key idea of the proof is to bound the reduction in the weight of the noisy mismatch vector Z through each step of the algorithm, and to show that when the weight of Z is reduced, the weight of the residual error is also subsequently reduced.There is a technical challenge to this idea however: there is no direct relation between the weight of Z and the error weight.
To bridge these two objects, we define the notion of an ideal mismatch vector Z (see Eq. ( 4) below), which is equal to Z when there is no measurement noise.Since the mismatch Z only captures the portion of the error which cannot be removed using independent local corrections, we must first "pre-process" the error by making any possible local corrections (see Eq. ( 5) below).This establishes a direct connection between Z and the "pre-processed" error e 0 (see Lemma 4.7) and our analysis will be built upon this connection.
We show that if Z is decomposable into local corrections by Algorithm 2, then most of these correction sets will also reduce the weight of Z (Lemma 4.13).This in turn allows us to relate the weights of Z and Z. Finally, we show that if the qubit and measurement error weights are bounded, the ideal mismatch vector Z always admits the desired decomposition into local correction sets (Lemma 4.15).These lemmas allow us prove our main result (Theorem 4.17): as the weight of Z decreases throughout the steps of the algorithm, the residual error weight must also decrease.The analysis of the parallel decoder then builds upon this bound, with the additional requirement of showing that the decomposition of Z into local corrections must be essentially disjoint (Lemma 4.18).
In the remainder of this section, we set up notation and provide some preliminary results used in the proofs of Theorems 4.17 and 4.20.We first define quantities relating to the states of the decoders.Given the local structure of the quantum Tanner codes, it will be more convenient to bound the size of the syndrome noise in terms of its vertex support.Given the noisy syndrome σ = H Z e + D, let σv denote the restriction of σ to the checks associated with the vertex v.For each vertex v ∈ V 1 , the decoder finds a locally minimal correction εv such that σ v (ε v ) = σv .In the event that no local correction εv exists for σv , we may define σv arbitrarily.
In our case, we will simply define εv = 0 by convention.If ε v is the locally minimal correction associated with the noiseless syndrome σ v , then we can decompose εv into "noiseless" and "noisy" parts as where ε v (D) is defined by ε v (D) = εv − ε v .Note that ε v (D) will be non-zero only when D has non-zero support on v.
The full noisy mismatch vector initialized by the decoder is given by Z = It will likewise be convenient to split the mismatch into a noiseless and a noisy part, defined by so that Z = Z + Z N .We will also need the restrictions of these vectors onto the vertices of V 01 , which we define as The key idea of the proof is to pre-process the error using Z01 , and apply the local corrections x v step by step.Specifically, we define the initial pre-processed error ẽ0 , and the "noiseless" pre-processed error e 0 , by ẽ0 = e + Z01 = e + v∈V 01 εv , e 0 = e + Z 01 = ẽ0 + Z 01 N .For the purpose of our proof, we consider the vector ẽ0 as the initial error state of the algorithm, and Z0 = Z as the initial mismatch.Note that in practice it does not matter at what point in the decoding procedure the set Z01 is flipped.The pre-processing is only introduced as a convenience in our proof in order to relate the weight of e to the weight of Z.The original algorithms considered in Ref. [33] involve a "post-processing" step instead, where Z01 is applied at the very end rather than the beginning.Since the sets of qubits flipped are ultimately the same in either case, the results here hold without modification.
The core loop of the decoding algorithm finds, at each step i, some local codeword Having found a codeword which satisfies (6), we update the error and the mismatch vectors by ẽi = ẽi−1 + f i , and Zi = Zi−1 + x i , where the flip-set f i ⊆ Q(v i ) is defined by Likewise, we can define the associated "noiseless" error and mismatch at each step by N are determined entirely by the syndrome noise D and initial error e, and are constant through the decoding process.
In the presence of measurement errors, it is no longer true that the noisy mismatch Z can be decomposed into a sum of local codewords. 3As such, some care must be taken in characterizing what exactly we mean by a "mismatch".This is captured by the definition below. 3In the case of perfect syndrome measurements, we have where rv + cv is the codeword that the local error is corrected to: ev + εv = rv + cv.This decomposition no longer holds in the presence of imperfect measurements.Definition 4.4.A mismatch vector is any Z ∈ F Q 2 that can be decomposed as Z = C 0 +C 1 +R 0 +R 1 , where e., a mismatch vector is an element in the span of local codewords C ⊥ 1 .Here, we define i = 1 − i for convenience.

The division of Z into local codewords of the form
Any given mismatch vector Z may have many distinct decompositions.Given any decomposition, we define its weight by where ∥C i ∥ and ∥R i ∥ denote the number of non-zero columns and rows, respectively, present in C i and R i .Note that the weight is well-defined since distinct local codewords c v ⊆ C i and r v ⊆ R i are disjoint.We then define the norm of a mismatch to be ∥Z∥ = min Decompositions such that wt(C 0 , C 1 , R 0 , R 1 ) = ∥Z∥ are called minimal weight decompositions for Z.
Note that technically the vector Z which we call the noisy mismatch vector is not a mismatch vector at all as defined by Definition 4.4.Nevertheless, we will continue to call Z the noisy mismatch since there is little chance of confusion.The noiseless part Z is a genuine mismatch vector by definition.The properties of the noiseless mismatch Z are characterized by the following lemma from Ref. [33].
Lemma 4.5 (Lemma 17 in Ref. [33]).Let e ∈ F Q 2 be an error and let ε v be a local minimal correction for e v at every vertex Then Z is a mismatch vector which satisfies The main purpose of pre-processing in our proof is that the noiseless pre-processed error e 0 and the noiseless mismatch Z 0 can be easily related through the following property.
Definition 4.6.Let e ∈ F Q 2 be an error.We say that the error is V ij -weighted if σ v (e) = 0 for all v ∈ V ij .Given a V ij -weighted error e, we say that a mismatch vector Z is associated with e if σ v (Z) = σ v (e) for all v ∈ V ij .
Lemma 4.7.The quantity e 0 is a V 10 -weighted error and Z 0 = Z is a mismatch vector associated with e 0 .
Proof.First, we show that Z is a mismatch vector.Note that Z is the sum of local minimal corrections ε v to the error e, i.e., where the e v terms cancel since each face occurs exactly twice in the sum above.Next, we show that e 0 is V 10 -weighted.We have Note that the terms in the latter sum are disjoint for distinct vertices v, v ′ ∈ V 01 .It follows that the restriction of e 0 to a vertex v ∈ V 01 is given by which has zero syndrome.Finally we show that Z is associated with e 0 .The restriction of e 0 to a vertex v ∈ V 10 is given by Likewise, the restriction of Z to v ∈ V 10 is given by It follows that which shows that Z is associated with e 0 .
The notion of e i being a V 10 -weighted error is invariant as the decoder proceeds, i.e., if e i is initially V 10 -weighted then it remains so.Moreover, if Z was initially a mismatch associated with e 0 then Z i remains associated with e i throughout all steps i of the decoder.
Lemma 4.8.Let Z be a weighted mismatch vector associated with a V 10 -weighted error e.Let to be the associated flip set.Then e + f is again a V 10 -weighted error and Z + x is an associated mismatch vector.
Proof.It is clear that Z +x is a mismatch vector since Z was one and we add a single C ⊥ 1 codeword.
We first show that e + f remains V 10 -weighted.Clearly e + f is V 10 -weighted if v ∈ V 10 or v ∈ V 01 since we either add nothing, or a local codeword to a V 01 vertex.Now suppose that v ∈ V 00 so that f = r.We can decompose r into r = r 1 + • • • + r k , where each r i is a local codeword supported on a single row, which we can assume to be indexed by the edge (v, u i ) for some u i ∈ V 01 .The syndrome of e + r on a vertex u ∈ V 01 is therefore given by σ u (e + f ) = σ u (e) u ̸ = u i for all i, σ u (e + r i ) u = u i for some i.
In either case, we have σ u (e + f ) = 0 so that e + f is V 10 -weighted.The case where v ∈ V 11 is analogous, taking f = c and making a similar decomposition.
Finally, we show that Z + x is associated with e + f .Let us write where σ u (Z) = σ u (e) for all u ∈ V 10 .If v ∈ V 10 then there is nothing to show since all syndromes are unchanged.If v ∈ V 01 then define Since (e + x) u = e u + Q(u) ∩ x, we see that ε ′ u has the same syndrome as (e + f ) u .
Lastly, suppose v ∈ V 00 , with the V 11 case being analogous.Let f = r.Note that ε ′ u = ε u and (e + r) u = e u for all u ∈ V 10 not adjacent to v. Therefore it suffices to consider u ∈ N (v).In this case, Q(u) ∩ c is just the column of c labeled by the edge (u, v) and so Q(u) ∩ c is a local codeword.Therefore σ u (c) = 0.It follows that for all u ∈ N (v).Therefore σ u (Z + x) = σ u (e + f ) for all u ∈ V 10 and so Z + x is associated with e + f .Lemma 4.7 and Lemma 4.8 show that Z i is a mismatch vector associated with the V 10 -weighted error e i for all i.We further cite the following lemma from Ref. [33], which gives a sufficient condition for the existence of good local corrections.This is the key to proving that in the noiseless case, the sequential and parallel decoders converge.
Definition 4.9.Let Z be a mismatch vector and let Z = C 0 + C 1 + R 0 + R 1 be a minimal decomposition for Z.We say that a vertex v ∈ V ij is active with respect this decomposition if Theorem 4.10 (Theorem 12 in Ref. [33]).Fix δ ∈ (0, 1).Let Z be a non-zero mismatch vector.If for all i, j ∈ {0, 1}, the set of active vertices S ij ⊆ V ij for a minimal decomposition of Z satisfies where d r denotes the relative distance of the local code, then there exists a non-zero

Sequential Decoder
To begin analyzing the sequential decoder with noisy input, the natural question to ask is that if the ideal mismatch Z can be decomposed by Algorithm 2 into F = {x i } t i=1 , how well do these local corrections x i decompose the noisy mismatch Z = Z + Z N ?The following two lemmas address this question.
Definition 4.11.Let Z be a mismatch vector.We say that Z is δ-decomposable if Algorithm 2 successfully returns a decomposition of Z when run with parameter δ, i.e., if Algorithm 2 halts with state Ẑ = 0. Lemma 4.12.Let Z be an δ-decomposable mismatch and let F = {x i } t i=1 denote the codewords returned by Algorithm 2 run with input Z and parameter δ.Then with Z = Z 0 .Note that since Algorithm 2 completely decomposes Z, we have Z t = 0 and For the decomposition with parameter δ, we have Together, we get the bounds Lemma 4.13.Let Z be a mismatch vector and let Z N ∈ F Q 2 be any vector.
for constants In particular, if Proof.This proof follows the idea of Lemma 5.1 in Ref. [42].Given any set y ∈ F Q 2 , we have For all y ∈ F \ F * , we have We then have where the last inequality follows from Lemma 4.12.On the other hand, we also have Combining these two inequalities, we get as desired.
Note that Lemma 4.13 will set an implicit bound of δ < 1/2 since we require ε − 2δ > 0 for the bound ( 7) to be non-trivial.
Suppose now that the noisy mismatch vector Z is given as input to Algorithm 1 with parameter ε, which terminates after T iterations.Let us denote the residual error by ẽT and its associated mismatch by ZT = Z T + Z N .If Z T is δ-decomposable, then Lemma 4.13 implies that |Z T | = O(Z N ).Namely, the sequential decoder terminates only when the mismatch noise Z N becomes significant.In the following lemma, we further relate the weight of the noiseless residual error e T with |Z T |.
Lemma 4.14 (Mismatch Correctness and Soundness).Let e be a V 10 -weighted error and let Z be an associated mismatch vector.Suppose that Z is δ-decomposable and that Then we have Proof.Let F = {x i } t i=1 denote the decomposition returned for Z by Algorithm 2 with parameter δ.Each x i is supported on the local view of some vertex v i and has the further decomposition into column and row codewords as First, we prove that e ∼ = Ĉ1 + R0 , where ∼ = denotes equivalence up to stabilizers.Let e 0 = e and define e i = e i−1 + f i where Note that by construction we have e t = e 0 + Ĉ1 + R0 .By Lemma 4.8, the errors e i are all V 10 -weighted, and the vector Z k = Z + k i=1 x i is a mismatch vector associated with e i at each step.It follows by the V 10 -weighting of e t that ∀v ∈ V 01 : σ v (e t ) = 0.
Since Z t = 0, it follows by the association of Z t and e t that ∀v ∈ V 10 : It follows that e t has zero syndrome.It remains to show that e t is a stabilizer, which we can do by bounding its weight.For each flip-set f i , we have where we use the robustness of the local code in the last inequality.Using Lemma 4.12, we then have It follows that Therefore e t ∼ = 0 and hence e ∼ = Ĉ1 + R0 .Finally, we have |Z|.
Now we show that, without surprise, Z T is δ-decomposable.Let us define the constants , and For the purposes of the parallel decoder, it will be convenient to consider a generalized mismatch decomposition procedure which initially starts the decomposition with some weight parameter ε and then switches to some other parameter ε ′ partway through (see Lemma 4.18).We state the generalized result below in Lemma 4.15, although we will only need the special case where ε = ε ′ for the analysis of the sequential decoder.
Lemma 4.15.Let e be an error and D a syndrome noise.Let Z ≡ Z + Z N denote the initial noisy mismatch vector assigned to e and D.
Let ε, ε ′ ∈ (0, 1) be constants such that ε ′ ≤ ε.Consider a modified Algorithm 2 which takes input Z and runs with parameter ε for the first t steps and then switches to parameter ε ′ until it halts at step T ≥ t.Let ZT ≡ Z T + Z N denote the final output of this process.
Proof.Consider the process of running the modified Algorithm 2 with input Z and parameter ε for t steps, and then switching the parameter to ε ′ until the algorithm finally halts at step T .Let {x 1 , . . ., x t } be local codewords obtained with parameter ε, and {x t+1 , . . ., x T } the codewords obtained with parameter ε ′ .Denoting Zi the mismatch vector at iteration i, we have We wish to show that Z T is δ-decomposable.Suppose that Algorithm 2 returns local codewords {y 1 , . . ., y K } when given input Z T with parameter δ.Let S T +k,ij denote a set of active vertices in V ij for the mismatch For all k ∈ [K], we have where the first inequality holds since there exists at least one non-zero row or column for each active vertex.By robustness of the local code, we have κ∆∥x i ∥ ≤ |x i |.Continuing the chain of inequalities, we have where the first inequality follows by robustness and the second by the fact that Using inequality (10), we get Inserting ( 13) into (12), we get where the last inequality follows by applying Lemma 4.5, together with the fact that ε v (D) can be non-zero only when v is in the support of D and hence Simplifying, we finally get where we use the fact that (2 − δ)/(1 − δ) ≤ 3 for δ ∈ (0, 1/2).It follows that if we have Then Proof.By Lemmas 4.7 and 4.8, the error e T is V 10 -weighted and Z T is an associated mismatch vector.Applying Lemma 4.14, it suffices to prove We have where the second inequality follows from (9).We then get where we use (13) in the second inequality.Next, we may assume without loss of generality that e is a reduced error.Then we have where we use the fact that ε v are minimum weight corrections in the inequality above and the fact that e u ∩ e v = ∅ for distinct vertices u, v ∈ V 01 in the second last equality.Following the same steps as from ( 14) to (15), we therefore get We can simplify the inequality above by noting that κ ≤ d r ≤ 1 [36].Then we have Therefore it suffices to require in order that inequality (16) holds.
Combining the inequalities, we obtain the main result for sequential decoder.
Theorem 4.17 (Main Theorem for the Sequential Decoder).Let e be an error and let D be a syndrome error.Suppose that Let σ = σ(e) + D. Then Algorithm 1 with input σ and parameter ε will output a correction f Proof.Suppose that Algorithm 1 with parameter ε terminates after T steps with output f .Let Z T denote the state of the mismatch after the algorithm terminates.By Lemma 4.15, Z T is δdecomposable.This allows us to apply Lemma 4.13, giving since the set F * must be empty when Algorithm 1 with parameter ε terminates.By Lemma 4.16, we get But we know Combining the inequalities ( 17), (18), and ( 19) finally gives Note that the restriction δ < 1/2, as required by Lemma 4.13, implies that (1 This completes our proof of the main theorem for the sequential decoder.

Parallel Decoder
The key idea in analyzing the parallel decoder is to compare the performance of one iteration of parallel decoding to that of a full execution of the sequential decoder.Our convention in this section will be that superscript indices will denote the parallel decoding iteration (always with parameter 1/2), while subscript indices will denote the sequential decoding iteration.For example, Z(k) j denotes the mismatch obtained after k iterations of parallel decoding and then j iterations of sequential decoding.
For convenience, we will fix some parameters in this section.Throughout, we will take ε = 1/2 for the parallel decoder.We will write A = A ε=1/2 and B = B ε=1/2 .Lemma 4.18.Let ε ′ ∈ (0, 1/6).Let Z(k) denote the current state of the (noisy) mismatch vector.Let Z(k) T denote the residual mismatch after running the sequential decoder with input Z(k) and parameter ε ′ .Then after one iteration of parallel decoding, the weight of the mismatch is reduced by at least Proof.The proof closely follows the ideas of Lemma 18 in Ref. [33].For ease of notation we write Z(k) as Z throughout this proof.Suppose that Algorithm 2 runs with input Z and parameter ε ′ returns local codewords {x i } T i=1 and residual mismatch ZT .Therefore we can write Z = T i=1 x i + ZT .
We will analyze the overlap among the sets x i , and argue that the parallel decoder's output will intersect non-trivially with the sequential decoder's output.Let us define the sets Note that the sets x ′ i are disjoint, and that they satisfy Next, we define the set of "good" indices G ⊆ [T ] such that i ∈ G if and only if We wish to bound the difference between Zj and Z′ j .Let us denote this difference by To bound the size of A j , we examine how the size of Z changes as we update it by adding codewords x j .Since the x j 's were obtained by running the decoder with parameter ε ′ , it follows that Referring to Figure 3, we have A j \ A j−1 = x j \ Zj−1 , and hence where F = {z v } is the collection of all local codewords found by the decoder in the current iteration.We now prove that for all j ∈ G, we have |x j ∩ u| ≥ c|x j | for some constant c > 0.
Fix some x j and let v denote its anchoring vertex.Let us write y = |x ′ j ∩ u|.First, let us show that we must have Suppose otherwise.Then let z v denote the codeword (possibly zero) that the parallel decoder assigns to vertex v.Note that we have z v ⊆ u by definition, as well as where Z denotes the current state of the noisy mismatch in the parallel decoder.By definition of u as the execution support of the decoder, the qubits of x ′ j \ u are untouched by the algorithm.Therefore, since x ′ j ⊆ Z, it follows that x ′ j \ u ⊆ Z and x ′ j \ u ⊆ Z + z v .Therefore we have The addition of x j to Z +z v therefore removes at least |x ′ j \u| ≥ 3 4 |x j | qubits from Z. Consequently, the addition of x j to Z + z v can add at most |x j |/4 qubits, so that Adding this inequality to (23), we get Similar to the argument above, the addition of x j to z v adds at least |x j \ z v | ≥ |x ′ j \ u| ≥ 3|x j |/4 qubits, and hence removes at most |x j |/4 qubits.Therefore this contradicts the assumption that z v is the local codeword selected by the decoder, since the decoder will choose to maximize the Hamming weight of its local codewords.It follows that we've established the inequality This then implies that for all j ∈ G, we have Since the x ′ j are disjoint, we get where the last inequality follows from (22).Finally, by the decoding criterion (23), the total decrease in mismatch weight is where we restore the superscript (k) in this last inequality for clarity.Now, as in the sequential case, we bound the weight of the residual mismatch by the weight of measurement noise.
Lemma 4. 19.Let e be an error and D be a syndrome noise.Let Z be the initial mismatch vector assigned to e and D. Let Z(k) denote the state of the mismatch vector after k iterations of parallel decoding.Let Z(k) T denote the residual mismatch vector obtained by running the sequential decoder with input Z(k) and parameter ε ′ .

Suppose that
Proof.Suppose that F = {x i } K i=1 are the codewords which have been found by the parallel decoder after k iterations.Note that we can equivalently consider the same sequence to be obtained by running the sequential decoder with parameter 1/2, i.e., we can consider Z(k) to be a state of the mismatch after K iterations of sequential decoding with parameter 1/2.It follows that Z(k) T is a mismatch obtained by first running the sequential decoder with input Z and parameter 1/2 for K iterations, and then switching to parameter ε ′ for the remaining iterations.
Applying Lemma 4.15 with ε = 1/2, our assumptions on |e| R and |D| V imply that Z (k) T is δdecomposable.Next, applying Lemma 4.13 (with ε ′ as ε), it follows that We then have For simplicity, we take ε ′ = 3δ in the following theorem.Note that this sets an upper bound on δ so that δ < 1/18.Summing this inequality over k gives Next, let ẽ(k) denote the state of the error after k iterations of parallel decoding.Let ẽ(k) T denote the state of the error after T additional iterations of sequential decoding with parameter ε ′ .Let us write ẽ(k) T = ẽ(k) + It remains to bound the weight of |ẽ (k) | R .We have ≤ e (k) In the above, the first inequality (26) follows from (19).Inequality (27) follows from (25) and the κ-product-expansion of the local code.Inequality (28) follows from the fact that each local codeword x i satisfies the decoding condition with parameter ε ′ .Finally, inequality (29) follows from (24).

DISCUSSION
In our article, we have shown that quantum Tanner codes admit single-shot QEC.Given information from a single round of noisy measurements, the mismatch decomposition decoder [33] is able to output a correction that is close to the data error that occurred.For a variety of noise models, including adversarial or stochastic noise, the single-shot decoder is able to maintain the encoded quantum information for up to an exponential number of correction rounds.The parallelized version of the decoder can be run in constant time while keeping the residual error small.During readout, a logarithmic number of iterations suffices to recover the logical information.
One may also ask about the possibility of single-shot QEC with other decoders for good QLDPC codes.Due to the close connection between the decoders analyzed here and the potential-based decoder for quantum Tanner codes in Ref. [22] (for example, the ability to map between candidate flip sets for both types of decoders), a corollary of the proofs presented here is that the potentialbased decoder also has the single-shot property.Likewise, under the mapping of errors shown in Ref. [21], the decoders considered here are applicable to the original good QLDPC codes by Panteleev and Kalachev [17].Our analysis does not straightforwardly carry over to the code and decoder proposed in Ref. [19], and it remains to be seen whether that construction also admits single-shot decoding.
We further remark that all known constructions of asymptotically good QLDPC codes admit a property called small-set (co)boundary expansion [43], which in the case of quantum Tanner codes, was used to prove the No Low-Energy Trivial States (NLTS) conjecture (see Property 1 of reference [44]).Small-set (co)boundary expansion is also equivalent to the notion of soundness [45], which lower bounds the syndrome weight by some function of reduced error weight.Indeed, soundness is a strong indication of single-shot decodability.Similarly, quantum locally testable codes [46][47][48][49][50] admit analogous soundness properties, although decoders for such codes are unexplored.Note that in our proof, what we needed was a notion of soundness for the mismatch vector (see Lemma 4.14), which is distinct from the usual notion of soundness for the syndrome.The weight of the mismatch is in general incomparable to the weight of the syndrome, so the precise relation between these two definitions of soundness is not well understood.
In conclusion, our results can be viewed as a step toward making general QLDPC codes more practical.While many challenges still remain, there have been promising developments in this direction [41,[51][52][53].We believe that quantum LDPC codes, similar to classical LDPC codes, will constitute the gold standard for future quantum telecommunication technologies and form the backbone of resource-efficient quantum fault-tolerant protocols.

Theorem 1 . 4 (
Informal Statement of Theorem 3.5).Consider an (α, β)-single-shot decoder and multiple rounds of errors (e i , D i ) for i = 1, • • • , M .For any c > 0, there exists a constant C * > 0 such that if max(|e i |, |D i |) ≤ C * n for all i, then the final residual error e ′ M satisfies |e ′ M | R ≤ cn.
where H A and H B are the parity check matrices of C A and C B , respectively.

FIG. 1 .
FIG. 1.The local structure of the left-right Cayley complex around a vertex labelled by g ∈ G.The incident faces Q(v) has a natural bijection with A × B. As examples, the red and blue faces in the complex are mapped to the squares of the same colors in the matrix given by A × B.

Definition 3 . 2 .
Let C be an n-qubit CSS code and e ∈ F n 2 be a Pauli X error.The stabilizer-reduced weight |e| R of e is defined as the weight of the smallest error equivalent to e up to the addition of stabilizers of C, i.e., |e| R = min e ′ ∈C ⊥ X |e + e ′ |.The stabilizer-reduced weight of a Pauli Z error is defined analogously.

4 .
PROOFS OF SINGLE-SHOT DECODING OF QUANTUM TANNER CODES4.1.Decoding AlgorithmsWe consider the decoding problem for quantum Tanner codes with parameters as in Theorem 2.3.We first provide an overview of how the decoder works.As before, we will work exclusively with X-type errors, with Z-errors being analogous.Suppose that the code state experiences data error e, and the measurements experience syndrome error D. The decoder is consequently given as input the noisy syndrome σ = σ + D = H Z e + D. Due to the structure of the code, the global syndrome σ can equivalently be viewed as a set of noisy local syndromes {σ v } v∈V 1 , where σv denotes the restriction of σ to the checks associated with the local code C ⊥ 1 at vertex v.At each V 1 vertex, the decoder computes a minimal weight correction εv ⊆ Q(v) based on the local syndrome σv , i.e, εv = argmin{|y| : y

Definition 4 . 3 .
Given a quantum Tanner code and a syndrome noise D, let us define D v to be the restriction of D to the set of stabilizer generators associated with vertex v.We define the vertex support of D to be the set of all vertices such that D v ̸ = 0. We denote the size of the vertex support by |D| V .Note that we have ∆ −2 |D| ≤ r −1 |D| ≤ |D| V ≤ |D|, where r is the number of stabilizer generators associated with the local code.
where {f i } T i=1 are the associated flip-sets with parameter ε ′ .It follows from Lemma 4.8 that e(k)T is V 10 -weighted with associated mismatch Z (k) T .Lemma 4.16 then implies that |e (k) where k X and k Z are the dimensions of C X and C Z , respectively.The distance is d = min(d X , d Z ), whered X = min x∈C Z \C ⊥ X |x| and d Z = min x∈C X \C ⊥ Z |x|.A quantum code of distance d can protect against any unknown error of weight less than d/2.A quantum code C ⊆ (C 2 ) ⊗n of dimension k and distance d is said to be an[[n, k, d]] code.A family of CSS codes is said to be low-density paritycheck (LDPC) if H X and H Z are sparse, i.e., have at most a constant number of non-zero entries in every column and row.
then the active vertex condition of Theorem 4.10 is always satisfied so that Algorithm 2 must be able to completely decompose Z T .