Byzantine setunion consensus using efficient set reconciliation
Abstract
Applications of secure multiparty computation such as certain electronic voting or auction protocols require Byzantine agreement on large sets of elements. Implementations proposed in the literature so far have relied on state machine replication and reach agreement on each individual set element in sequence.
We introduce setunion consensus, a specialization of Byzantine consensus that reaches agreement over whole sets. This primitive admits an efficient and simple implementation by the composition of Eppstein’s set reconciliation protocol with BenOr’s ByzConsensus protocol.
A free software implementation of this construction is available in GNUnet. Experimental results indicate that our approach results in an efficient protocol for very large sets, especially in the absence of Byzantine faults. We show the versatility of setunion consensus by using it to implement distributed key generation, ballot collection, and cooperative decryption for an electronic voting protocol implemented in GNUnet.
This is a revised and extended version of a paper published under the same title at ARES 2016.
Keywords
Byzantine agreement Secure multiparty computation Complexity1 Introduction
Byzantine consensus is a fundamental building block for faulttolerant distributed systems. It allows a group of peers to reach agreement on some value, even if a fraction of the peers are controlled by an active adversary. Theoryoriented work on Byzantine consensus often focuses on finding a single agreement on a binary flag or bit string [29]. More recent approaches for practical applications are mainly based on state machine replication (SMR), wherein peers agree on a sequence of state machine transitions. State machine replication makes it relatively easy to lift existing, nonfaulttolerant services to a Byzantine faulttolerant implementation [15]. Each request from a client triggers a state transition in the replicated state machine that provides the service.
A major shortcoming of SMR is that all requests to the service need to be individually agreed upon in sequence by the replica peers of the state machine. This is undesireable since in unoptimized SMR protocols, such as PBFT [15], a single transition requires O(n ^{2}) messages to be exchanged for n replicas. Some implementations [35] try to address this inefficiency by optimistically processing requests and falling back to individual Byzantine agreements only when Byzantine behavior is detected. In practice, this leads to very complex implementations whose correctness is hard to verify and that have weak progress guarantees [16].
The canonical example for a service where this inefficiency becomes apparent is the aggregation of values submitted by clients into a set. This scenario is relevant for the implementation of secure multiparty computation protocols such as electronic voting [17], where ballots must be collected, and auctions [12], where bids must be collected. A direct implementation that reaches agreement on a set of m elements with SMR requires m sequential agreements, each consisting of O(n ^{2}) messages.
We introduce Byzantine setunion consensus (BSC) as an alternative communication primitive that allows this aggregation to be implemented more efficiently. In order to implement the set aggregation service described above, the peers first reconcile their sets using an efficient set reconciliation protocol that is not faulttolerant but where the complexity is bounded even in the case of failures. Then, they use a variant of ByzConsensus [9] to reach Byzantine agreement on the union.
We assume a partially synchronous communication model, where nonfaulty peers are guaranteed to successfully receive values transmitted by other nonfaulty peers within an existing but unknown finite time bound [22]. Peers communicate over pairwise channels that are authenticated. Message delivery is reliable (i.e., messages arrive uncorrupted and in the right order) but the receipt of messages may be delayed. We make the same assumption as Castro and Liskov [14, 15] about this delay, namely that it does not grow faster than some (usually exponential) function of wall clock time. We assume a computationally unbounded adversary that can corrupt at most t=⌈n/3⌉−1 peers creating Byzantine faults. The adversary is static, that is the set of corrupted peers is fixed before the protocol starts, but this set is not available to the correct peers. The actual number of faulty peers is denoted by f, with f≤t.
The BSC protocol has message complexity O(m n+n ^{2}) when no peers show Byzantine behavior. When f peers show Byzantine behavior, the message complexity is O(m n f+k f n ^{2}), where k is the number of valid set elements exclusively available to the adversary. We will show how k can be bounded for common practical applications, since in the general case k is only bounded by the bandwidth available to the adversary. In practice, we expect kf to be significantly smaller than m. Thus, O(m n f+k f n ^{2}) is an improvement over using SMRPBFT which would have complexity O(m n ^{2}).
We have created an implementation of the BSC protocol by combining BenOr’s protocol for Byzantine consensus [9] with a bounded variant of Eppstein’s protocol for efficient set reconciliation [23]. We demonstrate the practical applicability of our resulting abstraction by using BSC to implement distributed key generation, ballot collection, and cooperative decryption from the CramerGennaroSchoenmakers remote electronic voting scheme [17] in the GNUnet framework. Our experimental results show that the behavior of the implementation matches our predictions from theory.

The introduction of Byzantine SetUnion Consensus (BSC) with Byzantine Eppstein Set Reconciliation

The analysis and proof of correctness of Byzantine SetUnion Consensus

An implementation and experimental evaluation of the protocol

A discussion of practical applications to secure multiparty computation
2 Background
The Byzantine consensus problem [37] is a generalization of the consensus problem where the peers that need to reach agreement on some value might also exhibit Byzantine faults.

Agreement: If peers P _{ i }, P _{ j } are correct, then v _{ i }=v _{ j }

Termination: The protocol terminates in a finite number of steps

Validity: If all correct peers have the same input value \(\tilde v\), then all correct peers decide on \(\tilde v\)
Some definitions of the consensus problem also include strong validity, which requires the value that is agreed upon to be the initial value of some correct peer [45]. The consensus protocol presented in this paper does not offer strong validity; in fact, for a setunion operation, this is not exactly desirable as the goal is to have all peers agree a union of all of the original sets, not on some peer’s initial subset.
2.1 The FLP impossibility result
A fundamental theoretical result (often called FLP impossibility for the initials of the authors) states, informally, that no deterministic protocol can solve the consensus problem in the asynchronous communication model, even in the presence of only one crash fault [28].

Common coins: Some protocols introduce a shared source of randomness that the adversary cannot predict or bias. This avoids the FLP impossibility result, since the protocol is not deterministic anymore. In practice, these protocols are very complex and often use variants of secretsharing and weaker forms of Byzantine agreement to implement the common coin [24, 25, 43]. Implementing a common coin oracle resilient against an active adversary is nontrivial and usually required extra assumptions such as a trusted dealer in the startup phase [13] or shared memory [6]. Recent designs to implement a Byzantine faulttolerant biasresistant public random generator only scale to hundreds of participants and still have relatively high failure rates (reported at 0.08% for and adversary power bounded at \(\frac {1}{3}\) and 32 participants) [53].

Failure oracles: Approaches based on unreliable failure detectors [33] augment the model with oracles for the detection of faulty nodes. Much care has to be taken not to violate correctness of the protocol by classifying too many correct peers as faulty; this problem is present in early systems such as Rampart [50] and SecureRing [34] as noted by Castro and Liskov [14, 15]. While the theory of failure detectors is quite established for the nonByzantine case, it is not clear whether they are still useful in the presence of Byzantine faults.

Partial synchrony: A model where a bound on the message delay or clock shift exists but is unknown or is known but only holds from an unknown future point in time is called partial synchrony. The FLP result does not hold in this model [22].

Minimal synchrony: The definition of synchrony used by the FLP impossibility result can be split into three types of synchrony: Processor synchrony, communication synchrony and and message ordering synchrony. Dolev et al. [21] show that consensus is still possible if only certain subsets of these three synchrony assumptions are fulfilled.
This work follows the path of [22] in relaxing the full asynchrony assumption behind the FLP impossibility result.
2.2 Byzantine consensus in the partially synchronous model
The protocols presented in this paper operate within the constraints of the partially synchronous model, where participants have some approximate information about time.
A fundamental result is that no Byzantine consensus protocol with n peers can support ⌈n/3⌉ or more Byzantine faults in the partially synchronous model [22].
Early attempts at implementing Byzantine consensus with state machine replication are SecureRing [34] and Rampart [50]. A popular design in the partially synchronous model is Castro and Liskov’s Practical Byzantine Fault Tolerance (PBFT) [14, 15]. PBFT uses a leader to coordinate peers (called replicas in BPFT terminology). When replicas detect that the leader is faulty, they run a leaderelection protocol to appoint a new leader.
PBFT guarantees progress as long as the message delay does not grow indefinitely for some fixed growth function^{2}. The approach taken by BPFT (and several derived protocols) has several problems [16]: In practice, malicious participants are able to slow down the system significantly. When facing an adversarial scheduler that violates PBFT’s weak synchrony assumption, PBFT can fail to make progress entirely [40].
Some more recent Byzantine state machine replication protocols such as Q/U [3] or Zyzzyva [35] have less overhead per request since they optimize for the nonByzantine case. This comes, however, often at the expense of robustness in the presence of Byzantine faults [16], not to mention that correctness proofs for the respective protocols and the implementation of state machine replication are notoriously difficult [8].
2.3 Gradecast
A key building block for our protocol is Feldman’s Gradecast protocol [24]. In contrast to an unreliable broadcast, Gradecast provides correctness properties to the receivers, even if the leader is exhibiting Byzantine faults.
 1.
LEAD: If i=L, send the input value v _{ L } to \(\mathcal {P}\)
 2.
ECHO: Send the value received in LEAD to \(\mathcal {P}\)
 3.
CONFIRM: If a common value \(\overline {v}\) was received at least n−t times in round ECHO, send \(\overline {v}\) to \(\mathcal {P}\). Otherwise, send nothing

If some \(\hat {v}\) was received at least n−t times in CONFIRM, output \(\langle \hat {v},2 \rangle \)

Otherwise, if some \(\hat {v}\) was received at least t+1 times in CONFIRM, output \(\langle \hat {v},1 \rangle \)

Otherwise, output 〈⊥,0〉. Here, ⊥ denotes a special value that indicates the absence of a meaningful value
 1.
If c _{ i }≥1 then \(\hat {v}_{i} = \hat {v}_{j}\) for correct P _{ i } and P _{ j }
 2.
If P _{ L } is correct, then c _{ i }=2 and \(\hat {v}_{i}=v_{L}\) for correct P _{ i }
 3.
c _{ i }−c _{ j }≤1 for correct P _{ i } and P _{ j }
When a correct peer P _{ i } receives a Gradecast with confidence 2, it can deduce that all other peers received the same message, but some other peers might have only received it with a confidence of 1. Receiving a Gradecast with confidence 1 also guarantees that all other correct peers received the same message. However, it indicates that the leader behaved incorrectly. No assumption can be made about the confidence of other peers. Receiving a Gradecast with confidence 0 indicates that the leader behaved incorrectly and, crucially, that all other correct peers know that the leader behaved incorrectly.
A simple counting argument proves that the above protocol satisfies the three Gradecast properties [24].
2.4 ByzConsensus
ByzConsensus [9] uses Gradecast to implement a consensus protocol for simple values. Each peer begins with a starting value \(s^{(1)}_{i}\) and the list of all n participants \(\mathcal {P}\). Each peer also starts with an empty blacklist of corrupted peers. If a peer is ever blacklisted, it is henceforth excluded from the protocol. In ByzConsensus, Gradecast is used to force corrupt peers to either expose themselves as faulty—and consequently be excluded—by gradecasting a value with low confidence, or to follow the protocol and allow all peers to reach agreement.
ByzConsensus consists of at most f+1 sequentially executed superrounds r∈1…f+1 where f≤t. In each superround, each peer leads a Gradecast using their candidate value \(s^{(r)}_{i}\); these n Gradecasts can be executed in parallel. Leaders where the Gradecast results in a confidence of less than 2 are put on the blacklist. Recall that different correct peers might receive a Gradecast with different confidence; thus, peers do not necessarily agree on the blacklist.
At the end of each superround, each peer computes a new candidate value \(s^{(r+1)}_{i}\) using the value that was received most often from the Gradecasts with a confidence of as least 1. If \(s^{(r)}_{i}\) was received more than n−t times, then r=f and the next round is the last round.
If the final candidate value does not receive a majority of at least 2t+1 among the n Gradecasts, or if the blacklist has more than t entries, then the protocol failed: either more than t faults happened or, in the partially synchronous model, correct peers did not receive a message within the designated round due to the delayed delivery.
ByzConsensus has message complexity O(f n ^{3}). While the asymptotic message complexity is obviously worse than the O(n ^{2}) of PBFT, there is a way to use set reconciliation to benefit from the parallelism of the Gradecast rounds and thereby reduce the complexity to O(f n ^{2}).
2.5 Set reconciliation
The goal of set reconciliation is to identify the differences between two large sets, say S _{ a } and S _{ b }, that are stored on two different machines in a network. A simple but inefficient solution would be to transmit the smaller of the two sets and let the receiver compute and announce the difference. Research has thus focused on protocols that are more efficient than this naive approach with respect to the amount of data that needs to be communicated when the sets S _{ a } and S _{ b } are large, but their symmetric difference S _{ a }⊕S _{ b } is small.
An early attempt to efficiently reconcile sets [41] was based on representing sets by their characteristic polynomial over a finite field. Conceptually, dividing the characteristic polynomials of two sets cancels out the common elements, leaving only the set difference. The characteristic polynomials are transmitted as a sequence of sampling points, where the number of sampling points is proportional to the size of the symmetric difference of the sets S _{ a } and S _{ b }. The number of sampling points can be approximated with an upper bound or increased on the fly should a peer be unable to interpolate a polynomial. While theoretically elegant, the protocol is not efficient in practice. The computational complexity of the polynomial interpolation grows as O(S _{ a }⊕S _{ b }^{3}) and uses rather expensive arithmetic operations over large finite fields.
A practical protocol was first proposed by Eppstein et al. in 2011 [23]. It is based on invertible Bloom filters (IBFs), a probabilistic data structure that is related to Bloom filters [11], and stratas for difference estimation.
2.5.1 Invertible Bloom filters
An IBF is a constantsize data structure that supports four basic operations: insert, delete, decode, and subtract.
Insert and delete operations are commutative operations encoding a key that uniquely identifies a set element, typically derived from the element via a hash function.
The decode operation can be used to extract some or all of the updates, returning the key and the sign of the operation, which is either insert or delete. Since the data structure uses constant space, decoding cannot always succeed. Decoding is a probabilistic operation that is more likely to succeed when the IBF is sparse, which is when the number of encoded operations (excluding the operations that canceled each other out) is small. The decoding process can also be partially successful, if some elements could be extracted but the remaining IBF is nonempty. Extracting an update by decoding an IBF is only possible if the key was recorded only once in the IBF. However, storing a deletion or insertion of the same key twice or more (not counting operations that canceled each other out) makes both updates impossible to decode.
IBFs of the same size can also be subtracted from each other. When subtracting IBF_{ b } from IBF_{ a }, the resulting structure IBF_{ c }:=IBF_{ a }−IBF_{ b } contains all insertions and deletions from IBF_{ a }, and insertions from IBF_{ b } are recorded as deletions in IBF_{ c } and deletions from IBF_{ b } are recorded as insertions in IBF_{ c }. Effectively, the IBF subtraction allows to compute the difference between two sets simply by encoding each set as an IBF using only insertion operations.

A signed counter that handles overflow via wraparound. Recording an insertion or deletion adds −1 or +1 to the counter, respectively.

An ⊕sum^{3}, called the keySum, over the keys that identify set the elements that were recorded in the bucket.

An ⊕sum, called the keyHashSum, over a the hash h(·) of each key that was recorded in the bucket.
As with ordinary Bloom filters, encoding an update in an IBF records the update in k different buckets of the IBF. The indices of buckets that record the update are derived via a kindependent hash functions from the key of the element that is subject of the update. We write Pos(x) for the set of array positions that correspond to the element key x.
Before we describe the decoding process, we introduce some terminology. A bucket is called a candidate bucket if its counter is −1 or +1, which might indicate that the keySum field contains the key of an element that was the subject of an update. Candidate buckets that contain the key of an element that was previously updated are called pure buckets. Candidate buckets are not necessarily pure buckets, since a candidate bucket could also result from, for example, first inserting an element key e _{1} and then deleting e _{2} when Pos(e _{1})∩Pos(e _{1})≠∅ and Pos(e _{1})≠Pos(e _{2}).
The keyHashSum provides a way to detect if a candidate bucket is not a pure bucket, namely when h(keySum)≠keyHashSum. The probability of classifying an impure bucket as pure with this method is dependent on the probability of a hash collision. Another method to check for an impure candidate bucket with index i is to check whether i∉Pos(keySum).
The decoding process then simply searches for buckets that are, with high probability, pure. When the count field of the bucket is 1, the key decoding procedure reports the key as “inserted” and exececutes a deletion operation with that key. When the count field is −1, the key is reported as “deleted” and subsequently an insertion operation is executed.
With a probability that increases with sparser IBFs, decoding one element may cause one or more other buckets to become pure, allowing the decoding to be repeated. If none of the buckets is pure, the IBF is undecodable, and a larger IBF must be used, or the reconciliation could fall back to the naive approach of sending the whole set.
The IBF decoding process is particularly suitable for reconciling large sets with small differences. When the symmetric difference between the sets is small enough compared to the size of the IBFs, the result IBF_{ c } of the subtraction can be decoded, since the common elements encoded in IBF_{ a } and IBF_{ b } cancel each other out. This makes it possible to obtain the elements of the symmetric difference, even when the IBFs that represent the full sets can not be decoded.
As long as the symmetric difference between the original sets S _{ a } and S _{ b } can be approximated well enough, IBFs can be used for set reconciliation by encoding S _{ a } in IBF_{ a } and S _{ b } in IBF_{ b }. One of the IBFs is sent over the network, the IBF_{ c }=IBF_{ a }−IBF_{ b } is computed and decoded. Should the decoding (partially) fail, the same procedure is repeated with larger IBFs.
2.5.2 Difference estimation with stratas
In order to select the initial size of the IBF appropriately for the set reconciliation protocol, one needs an estimate of the symmetric difference between the sets that are being reconciled. Eppstein et al. [23] describe a simple technique, called strata estimation, which is accurate for small differences. While Eppstein et al. suggest combining the strata estimator, with a minwise estimator, which is more accurate for large differences, our work only requires the strata estimators.
A strata estimator is an array of fixedsize IBFs. These fixedsize IBFs are called strata since each of them contains a sample of the whole set, with increased sampling probability towards inner strata. Similar to how two IBFs can be subtracted, strata estimators are subtracted by pairwise subtraction of the IBFs they consist of.
The set difference is estimated by having both peers encode their set in a strata estimator. One of the strata estimators is then sent over to the other peer, which subtracts the strata estimators from each other. With every IBF of the strata estimator that results from the subtraction, a decoding attempt is made. The number of successfully decoded elements in each stratum allows an estimate to be made on the set difference, which is then used to determine the size of the IBF for the actual set reconciliation.
3 Our approach
We now describe how to combine the previous approaches into a protocol for Byzantine faulttolerant set consensus. The goal of the adversary is to sabotage timely consensus among correct peers, e.g., by increasing message complexity or forcing timeouts.
A major difficulty with agreeing on a set of elements as a whole is that malicious peers can initially withhold elements from the correct peers and later send them only to a subset of the correct peers. This could possibly happen at a time when it is too late to reconcile the remaining difference caused by distributing these elements. We assume that the number of these elements that are initially known to the adversary but not to all correct peers is bounded by k, where k exists but is not necessarily known to the correct participants.
3.1 Definition
We now give a definition of setunion consensus that is motivated by practical applications to secure multiparty computation protocols such as electronic voting, which are discussed in more detail in the section “Application to SMC.”
Consider a set of n peers \(\mathcal {P} = \{P_{1},\ldots,P_{n}\}\). Fix some (possibly infinite) universe M of elements that can be represented by a bit string. Each peer P _{ i } has an initial set \(S^{(0)}_{i} \subseteq M\).
 1.
For any pair of correct peers P _{ i }, P _{ j }, it holds that S _{ i }=S _{ j }.
 2.
If P _{ i } is correct and \(e \in S^{0}_{i}\), then e∈S _{ i }.
 3.
The set S _{ i } is canonical, which is S _{ i }=R(S _{ i }).
The canonicalization function allows us to set an upper bound on the number of elements that can simultaneously be in a set. For example in electronic voting, canonicalization would remove malformed ballots and combine multiple different (encrypted) ballots submitted by the same voter into a single “invalid” ballot for that voter.
3.2 Byzantine setunion consensus (BSC) protocol
Recall that every peer P _{ i }, 0<i≤n starts with a set \(S^{(0)}_{i}\). The BSC protocol incorporates two subprotocols, bounded set reconciliation and lower bound agreement, and uses those to realize an efficient Byzantine faulttolerant variant of ByzConsensus. An existing generalization of IBFs to multiparty set reconciliation [42] based on network coding is not applicable to this problem, as it requires trusted intermediaries.
The basic problem solved by the two subprotocols is bounding the cost of Eppstein’s set reconciliation. Given a set size difference between two peers of k, the expected cost of Eppstein’s set reconciliation is O(k) if both participants are honest. However, we need to ensure that malicious peers cannot generally raise the complexity to O(m) where m is the size of the union.
For this, we use a bounded variant of Eppstein’s set reconciliation protocol, which is given a lower bound \(\mathcal {L}\) on the size of the set of elements shared by all honest participants. Given such a lower bound, the bounded set reconciliation protocol must detect faulty participants in \(O(k + (m  \mathcal {L}))\). We note that for \(\mathcal {L} = 0\), the bounded set reconciliation is still allowed to cost O(m).
3.2.1 Bounded set reconciliation
In bounded set reconciliation, we are thus concerned with modifications that ensure that a set reconciliation step between an honest and a faulty peer either succeeds after O(k) traffic or aborts notifying the honest peer that the other peer is faulty. While we use probabilities to detect faulty behavior, we note that suitable parameters can be used to ensure that falsepositives are rare, say 1:2^{128}, and thus as unlikely as successful bruteforce attacks against canonical cryptographic primitives, which BSC also assumes to be safe.
To begin with, to bound the complexity of Eppstein set reconciliation one needs to bound the number of iterations the protocol performs. Assuming honest peers, the initial strata estimation ensures that the IBFs will decode with high probability, resulting in Eppstein’s claim of singleround complexity. Given aggressive choices of the parameters to improve the balance between roundtrips and bandwidth consumption, decoding failures can happen with nonnegligible probability in practice. In this case, the process can simply be restarted using a different set of hash functions and an IBF doubled in size. This addresses issues caused by conservative choices for IBF sizes that optimize for the average case. What is critical is that the probability of such failures remains small enough that after if the number of rounds exceeds some constant, we can assert faulty behavior and overall remain within the O(k) bound assuming individual rounds are bounded by O(k).
Another problem with Eppstein’s original protocol related to aggressive parameter choices is that iterative decoding does not always have to end with an empty or an undecodable IBF. Specifically, the decoding step can sometimes decode a key that was never added to the IBF, simply because the two purity checks are also probabilistic. This is usually not an issue, as when a decoder requests the transmission of the element corresponding to improperly decoded key, the presumed element’s owner can indicate a decoding failure at that time. Here, another round of the protocol is unlikely to produce the same error and would again fix the problem. However, given reasonably short strings for the hashKeySum, it is actually even possible to obtain a looping IBF that spawns an infinite series of “successfully” decoded keys. Here, the implementor has to be careful to ensure that the iterated decoding algorithm terminates. Instead of mandating an excessively long hashKeySum to prevent this, it is in practice more efficient to handle this case by stopping the iteration and reporting the IBF as undecodable when the number of decoded keys exceeds a threshold proportional to the size of the IBF.
We also need to consider the bandwidth consumption of an individual round. To cause more than O(k) traffic, a malicious peer could produce strata that result in a huge initial symmetric difference. In this case, the initial size of the IBF may exceed O(k). We address this problem by not permitting the use of Eppstein’s method if the symmetric difference definitively exceeds \(\frac {n  \mathcal {L}}{2}\), where n is the smaller of the two set sizes.^{4} Instead, once the estimate of the symmetric difference substantially exceeds this threshold, the reconciliation algorithm falls back to sending the complete set. As this creates O(m) traffic, it must only be allowed under certain conditions.
First, we consider the case where the honest peer has the larger set. Here, the honest peer P _{ i } will only send its full set if the set difference is no larger than \(S_{i}  \mathcal {L}\), and otherwise report a fault. This ensures that a malicious peer cannot arbitrarily request the full set from honest peers.
Second, we consider the case where the honest peer P _{ i } is facing a faulty peer that claims to have a huge set. This can happen either directly from the strata estimator or after P _{ i } observes a constant number of successive IBF decoding failures.^{5} At this point, instead of passively accepting the transmission of elements, the receiver P _{ i } checks that a sufficient number of the elements received are not in S _{ i }. Let R be the stream of elements e received at any point in time. We assume that the sender is required to transmit the elements in a randomized order. Thus, if R∩S _{ i }−R∖S _{ i }≥128, P _{ i } can determine that the sender is faulty with probability 2^{128}:1, as the the \(\frac {n}{2}\)threshold for converting to complete set transmission ensures that for an honest sender less than half of the elements would be in S _{ i }.
Finally, we note that the individual insert, delete, decode, and subtract operations on the IBF are all constant time and that IBFs are also constant size. Thus, given a constant number of rounds and a bound on the bandwidth per round, we have implicitly assured that memory and CPU consumption of the bounded set reconciliation is also \(O(k + (m  \mathcal {L}))\).
3.2.2 Lower bound agreement
 (i)
All peers reconcile their initial set with each other, using pairwise bounded set reconciliation using a lower bound of \(\mathcal {L} = 0\).
 (ii)
All peers send their current set size to each other, and each peer P _{ i } sets ℓ _{ i } to the (t+1) smallest set size that P _{ i } received.
 (iii)
All peers again reconcile their sets with each other, using pairwise bounded set reconciliation.
The third step is necessary to ensure that every correct P _{ i } has at least ℓ _{ i } elements, since malicious peers could use the k elements initially withheld to force an honest peer’s set size below the (t+1) smallest set size. Thanks to the repetition, even if ℓ _{ i } is different for each peer, it is guaranteed that P _{ i } has at least ℓ _{ i } elements in common with every other good peer.
In subsequent set reconciliations, ℓ _{ i } can be used to bound the traffic that malicious peers are able to cause by falsely claiming to have a large number of elements missing. LBA itself has complexity O(n m f); initially, all malicious peers can once claim to have empty sets with all other peers. LBA ensures that for the remainder of the protocol, a correct peer with m _{ i } elements can stop sending elements to malicious peer P _{ M } after P _{ M } requested m _{ i }−ℓ _{ i }≤k elements by reducing the complexity of bounded set reconciliation with peer m _{ i } to O(k) using \(\mathcal {L} = \ell _{i}\).
3.2.3 Exact set agreement
After LBA, an exact set agreement is executed, where all peers reach Byzantine agreement for a superset of the set reached in LBA. The exact set agreement is implemented by executing a variant of ByzConsensus which instead of sending values reconciles sets.
 (i)
LEAD: If i=L, reconcile the input set V _{ L } with \(\mathcal {P}\).
 (ii)
ECHO: Reconcile the set received in LEAD with \(\mathcal {P}\).
 (iii)
CONFIRM: Let \(\mathcal {U}_{E}\) be the union of all sets received in the ECHO round, and N _{ E }(e) the number of times a single set element e was received.
If \(\bigvee _{e \in {\mathcal {U}_{E}}} t < N_{E}(e) < nt\), send ⊥ (where ⊥≠∅). Otherwise, send \(\mathcal {U}_{E}  \{e \mid N_{E}(e) \le t\}\) to \(\mathcal {P}\).

If \(\bigwedge _{e \in \mathcal {U}} N_{C}^{+}(e) \ge nt \lor N_{C}^{}(e) \ge nt\),
the output is \(\langle \{e \mid N_{C}^{+}(e) \ge nt\},2 \rangle \).

Otherwise, if \(\bigwedge _{e \in \mathcal {U}_{C}} N_{C}^{+}(e) > t \land N_{C}^{+}(e) \ge N_{C}^{}(e)\)
or \(\bigwedge _{e \in \mathcal {U}_{C}} N_{C}^{}(e) > t \land N_{C}^{}(e) > N_{C}^{+}(e)\),
the is output \(\langle \{e \mid N_{C}^{+}(e) > t \land N_{C}^{+}(e) \ge N_{C}^{}(e)\},1 \rangle \).

Otherwise, the output is 〈⊥,0〉.
Similar to ByzConsensus, the BSC consists of at most f+1 superrounds, where f≤t. Each peer P _{ i } starts with \(s^{(1)}_{i}\) as its current set. In sequential superrounds, all peers lead a Gradecast for their candidate set. Like in ByzConsensus, if P _{ i } receives a Gradecast with a confidence value that is not 2, then P _{ i } puts the leader of the Gradecast on its blacklist, and correct peers stop all communictation with peers on their blacklist.
At the end of each superround, peers update their candidate set as follows. Let n ^{′} be the number of leaders that gradecasted a set with a nonzero confidence. The new candidate set contains all set elements that were included in at least ⌈n ^{′}/2⌉ sets that were gradecasted with a nonzero confidence value. If all elements occur with a (n−t) majority, then the next round is the last round. The output of the consensus protocol is the candidate set after the last round—or failure if f>t.
We give a correctness proof that generalizes Feldman’s proof for Gradecast of single values [25, “The GNUnet framework” section].
Lemma 1
If two correct peers send sets A≠⊥ and B≠⊥ in CONFIRM, then A=B.
Proof
Proof by contradiction and counting argument. Assume w.l.o.g. that e∈A and e∉B. At least n−t peers must have echoed a set that includes e to the first peer. Suppose f of these peers were faulty, then at least n−t−f>t good peers included e in the ECHO transmission to the second peer. If e∉B, then t<N _{ E }(e)<n−t. In this case, an honest second peer must output B=⊥. Contradiction. □
Theorem 1
The generalization of Gradecast to sets satisfies the three Gradecast properties.
Proof

Property 1 (If c _{ i },c _{ j }≥1, then \(\hat {v}_{i} = \hat {v}_{j}\) for correct P _{ i } and P _{ j }): Assume w.l.o.g. that \(e \in \hat {V}_{i} \setminus \hat {V}_{j}\).
For \(e \in \hat {V}_{j}\), P _{ i } must have received e at least \(N_{C}^{+}(e) > t\) times in CONFIRM. Given f≤t failures, at least one honest peer must thus have included e in CONFIRM. According to Lemma 1, all n−f honest peers must either include e in CONFIRM or send ⊥.
Because ⊥ is not a set, this leaves at most all f≤t faulty peers that can send a set without e. But for \(e \notin \hat {V}_{j}\), we need \(N^{}_{C}(e) \ge t + 1\). Contradiction.

Property 2 (If P _{ L } is correct, then c _{ i }=2 and \(\hat {V}_{i}=\hat {V}_{L}\) for correct P _{ i }): All n−f≥n−t good peers ECHO and CONFIRM the same set. By the grading rules, they must output a confidence of 2.

Property 3 (c _{ i }−c _{ j }≤1 for correct P _{ i } and P _{ j }): Proof by contradiction. Assume w.l.o.g. c _{ i }=2 and c _{ j }=0, c _{ i }=2 implies that for each \(x \in \hat {V}_{i}\) at least n−t peers (and thus (n−t)−f≥t+1 correct peers) must have sent a set in CONFIRM that includes x. For any \(y \notin \hat {V}_{i}\), n−t peers (and thus (n−t)−f≥t+1 correct peers) must have sent a non ⊥ set in CONFIRM that excludes y.
Given c _{ j }=0, there must have been an element e such that \(N_{C}^{+}(e) \le t\) and \(N_{C}^{}(e) \le t\) for P _{ j }. However, we just derived that for all elements, either \(N_{C}^{+}(e) > t\) or \(N_{C}^{}(e) > t\). Contradiction.
Given the Gradecast properties for sets, the correctness argument given by BenOr [9] for the Byzantine consensus applies to BSC’s generalization to sets.
As described, the protocol has complexity O(m n f+f k n ^{3}). However, the n parallel set reconciliation rounds in each superround can be combined by tagging the set elements that are being reconciled in the LEAD, ECHO, and CONFIRM rounds with the respective leader L. Because LBA (via n−ℓ _{ i }≤k) and bounded set reconciliation limit mischief for the combined superround, each malicious peer can, as leader, once cause bounded set reconciliation during the ECHO round to alltoall transmit at most k extra elements, resulting in a total of O(f k n ^{2}) extra traffic over all f+1 rounds. Before exposing themselves this way, nonleading malicious peers can only cause O(f ^{2} k n) additional traffic during all ECHO rounds. Finally, malicious peers can also cause at most O(f k n ^{2}) traffic in the CONFIRM round. Thus, BSC has overall message complexity of O(m n f+f k n ^{2}).
4 Implementation
We implemented the BSC protocol in the SET and CONSENSUS services of GNUnet [1].
4.1 The GNUnet framework
GNUnet is composed of various components that run in separate operating system processes and communicate via message passing. Components that expose an interface to other components are called services in GNUnet. The main service used by our implementation is the CADET service, which offers pairwise authenticated endtoend encryption between all participants. CADET uses a variation of the Axolotl public key ratcheting scheme and doubleencrypts using both TwoFish and AES. [49] The resulting encryption is relatively expensive compared to the other operations, and thus dominates in terms of CPU consumption for the experiments.
4.2 Set reconciliation
Bounded set reconciliation is implemented in the SET service. The SET service provides a generic interface for set operations between two peers; the operations currently implemented are the IBFbased set reconciliation and set intersection [54].
In addition to the operationspecific protocols, the following aspects are handled generically (i.e. independent of the specific remote set operation) in the SET service: Local set operations Applications need to create sets and perform actions (iteration, insertion, deletion) on them locally. Concurrent modifications While a local set is in use in a network operation, the application may still continue to mutate that set. To allow this without interfering with concurrent network operations, changes are versioned. A network operation only sees the state of a set at the time the operation was started. Lazy copying Some applications building on the SET service—especially the CONSENSUS service described in the next section—manage many local sets that are large but only differ in a few elements. We optimize for this case by providing a lazy copy operation that returns a logical copy of the set without duplicating the sets in memory. Negotiating remote operations In a network operation, the involved peers have one of two roles: The acceptor, which waits for remote operation requests and accepts or rejects them and the initiator, which sends the request.
Our implementation estimates the initial difference between sets only using strata estimators as described by Eppstein [23]. However, we compress the strata estimator—which is 60 KB uncompressed—using gzip. The compression is highly effective at reducing bandwidth consumption due to the high probability of long runs of zeros or ones in the most sparse or most dense strata.
We also use a salt when deriving the bucket indices from the element keys. When the decoding of an IBF fails, the IBF size is doubled and the salt is changed. This prevents decoding failures in scenarios where keys map to the same bucket indices even modulo a power of two, where doubling the size of the IBF does not remove the collision.
4.3 Setunion consensus
To keep the description of the setunion consensus protocol in the previous section succinct, we merely stated that peers efficiently transmit sets using the reconciliation protocol. However, given that the receiving peer has usually many sets to reconcile against, an implementation needs to be careful to ensure that it scales to large sets as intended.
The key goal is to avoid duplicating full sets and to instead focus on the differences. New sets usually differ in only a few elements; thus, our implementation avoids copying entire sets. Instead, in the leader round, we just store the set of differences with a reference to the original set. In the ECHO and CONFIRM round, we also reconcile with respect to the set we received from the leader and not a peer’s current set. In the ECHO round, we only store one set and annotate each element to indicate which peer included or excluded that element. This also allows for a rather efficient computation of the set to determine the ⊥ result in the CONFIRM round.
4.4 Evaluating malicious behavior

SpamAlways: A malicious peer adds a constant number of additional elements in every reconciliation.

SpamLeader: A malicious peer adds a constant number of additional elements in reconciliations where the peer is the leader.

SpamEcho: A malicious peer adds a constant number of additional elements in echo rounds.

Idle: Malicious peers do not participate actively in the protocol, which amounts to a crash fault from the start of the protocol. This type of behavior is not interesting for the evaluation, but used to test the implementation with regards to timeouts and majority counting.
For the Spam* behaviors, two different variations are implemented. One of them (“*replace”) always generate new elements for every reconciliation. This is not typical for real applications where the number of stuffable elements ought to be limited by set canonicalization. However, this shows the performance impact in the worst case. The other variation (“*noreplace”) reuses the same set of additional elements for all reconciliations, which is more realistic for most cases. We did not implement adversarial behavior where elements are elided, since the resulting traffic is the same as for additional elements, and memory usage would only be reduced.
5 Experimental results
All of the experiments were run on a single machine with a 24core 2.30 GHz Intel Xeon E52630 CPU and GNUnet SVN revision 36765. We used the gnunetconsensusprofiler tool, which is based on GNUnet’s TESTBED service [55], to configure and launch multiple peers on the target system. We configured the profiler to emulate a network of peers connected in a clique topology (via loopback, without artificial latency). Elements for the set operations are randomly generated and always 64 bytes large.
Bandwidth consumption was measured using the statistics that GNUnet’s CADET service [49] provides. Processor time was measured using GNUnet’s resource reporting functionality, which uses the wait3 system call for that purpose.
5.1 Bounded set reconciliation
Figure 2 shows that bandwidth consumption does not grow linearly with the total set size, as long as the set size difference between the two peers is small. The logarithmic increase of the traffic with larger sets can be explained by the compression of strata estimators: the kth strata samples the set with probability 2^{−k }, and for small input sets, the strata tends to contain long runs of zeros that are more easily compressed.
5.2 Byzantine set consensus
For our experiments with the BSC implementation, all ordinary peers start with the same set of elements; different sets would only affect the alltoall union phase of the protocol which does pairwise set reconciliation, resulting in increased bandwidth and CPU consumption proportional to the set difference as shown in the previous section.
Figure 11 highlights how the different attack strategies impact the number of additional elements that were received during set reconciliations; the number of stuffed elements for the “SpamEcho” behavior is significantly larger than for “SpamLead”, since multiple ECHO rounds are executed for one LEAD round, and the number of stuffed elements is fixed per reconciliation. When malicious peers add extra elements during the LEAD round, the effect of that is amplified, since all correct receivers have to redistribute the additional elements in the ECHO/CONFIRM round. Even though adding elements in the LEAD round requires the least bandwidth from the leader the effect on traffic and latency is the largest (see Figures 9 and 10).
As expected, when the number of stuffed elements is limited to a fixed set, the effect on the performance is limited (“SpamAllnoreplace” in Figures 9, 10, and 11).
6 Opportunities for further improving BSC
We now discuss some of the key limitations of the current implementation and how it could be optimized further.
6.1 Extension to partial synchrony
The prototype used in the evaluation only works in the synchronous model. It would be trivial to extend it to the partially synchronous model with synchronous clocks by using the same construction as BPFT [15], namely retrying the protocol with larger round timeouts (usually doubled on each retry) when it did not succeed.
It might be worthwhile to further investigate the Byzantine round synchronization protocols discovered independently by Attya and Dolev [7] as well as Dwork, Lynch, and Stockmeyer [22]. Running a Byzantine clock synchronization protocol interleaved with consensus protocol might lead to a protocol with lower latency, since the timeouts are dynamically adjusted instead of being increased for each failed iteration.
6.2 Persistent data structures
Both the SET and CONSENSUS service have to store many variations of the same set when faulty peers elide or add elements. While the SET service API already supports lazy copying, the underlying implementation is inefficient and based on a log of changes per element with an associated version number. It might be possible to reduce memory usage and increase performance of the element storage by using data structures that are more well suited, such as the persistent data structures described by Okasaki [46].
6.3 Fast dissemination
Recall that in order to be included in the final set, an element must be sent to at least t+1 peers, so that at least one correct peer will receive the element. In applications of setunion consensus such as electronic voting, the effort to the client should be minimized, and thus in practice elements might be sent only to t+1 peers, which would lead to large initial symmetric differences between peers.
A possible optimization would be to add another dissemination round that only requires n log2n reconciliations to achieve perfect element distribution when only correct peers are present. The n ^{2} reconciliations that follow will consequently be more efficient, since no difference has to be reconciled when all peers are correct. In the presence of faulty peers, the optimization adds more overhead due to the additional dissemination round.
More concretely, in the additional dissemination round the peers reconcile with their 2^{ ℓ }th neighbor (for some arbitrary, fixed order on the peers) in the ℓth subround of the dissemination round. After ⌈log2⌉ of these subrounds, the elements are perfectly distributed as long as every peer passed along their current set correctly.
7 Application to SMC
Secure multiparty computation (SMC) is an area of cryptography that is concerned with protocols that allow a group of peers \(\mathcal {P} = P_{1},\dots,P_{n}\) to jointly compute a function y=f(x _{1},…,x _{ n }) over private input values x _{1},…,x _{ n } without using a trusted third party [32]. Each peer P _{ i } contributes its own input value x _{ i }, and during the course of the SMC protocol, P _{ i } ideally only learns the output y, but no additional information about the other peers’ input values. Applications of SMC include electronic voting, secure auctions and privacypreserving data mining.
SMC protocols often assume a threshold t<n on the amount of peers controlled by an adversary, which is typically either honestbutcurious (i.e., tries to learn as much information as possible but follows the protocol) or actively malicious. The actively malicious case mandates the availability of Byzantine consensus as a building block [51].^{6}
In practical applications, the inputs typically consist of sets of values that were given to the peers \(\mathcal {P}\) by external clients. In electronic voting protocols, the peers need to agree on the set of votes; with secure auctions, the peers need to agree on bids, and so on.
In this section, we focus on one practical problem, namely electronic voting. We show how BSC is useful at multiple stages of the protocol and discuss how our approach differs from existing solutions found in the literature.
7.1 Bulletin board for electronic voting
The bulletin board is a communication abstraction commonly used for electronic voting [10, 48]. It is a stateful, appendonly channel that participants of the election can post messages to. Participants of the election identify themselves with a public signing key and must sign all messages that they post to the bulletin board. The posted messages are publicly available to facilitate independent auditing of elections.
Existing work on electronic voting either does not provide a Byzantine faulttolerant bulletin board in the first place [4] and instead relies on trusted third parties or suggests the use of state machine replication [17].
Some of the bulletin board protocols surveyed by Peters [48] use threshold signatures to certify to the voter that the vote was accepted by a sufficiently large fraction of the peers that jointly provide the bulletin board service. With a naive approach, the message that certifies acceptance by t peers is the concatenation of the peers’ individual signatures and thus O(t) bits at large. Threshold signature schemes allow smaller signatures, but at the expense of a more complex protocol. Since the number of peers is typically not very large, a linear growth in t is acceptable, which makes the simple scheme sufficient for practical implementations.
It is easy to implement a variant of the bulletin board with setunion consensus. In contrast to traditional bulletin boards, this variant has phases, where posted messages are only visible after the group of peers have agreed that a phase is concluded. The concept of phases maps well to the requirements of existing voting protocols. Every phase is implemented with one setunion consensus execution. To guarantee that a message is posted to the bulletin board, it must be sent to at least one correct peer from the group of peers that jointly implements the bulletin board.
7.2 Distributed threshold key generation and cooperative decryption
Voting schemes as well as other secure multiparty computation protocols often rely on threshold cryptography [19]. The basic intuition behind threshold cryptography is that some operations—such as signing a message or decrypting a ciphertext—should only succeed if a large enough fraction of some group of peers cooperate. Typically, the public key of the threshold cryptosystem is publicly known, while the private key is not known by any entity but reconstructible from the shares that are distributed among the participants, for example with Shamir’s secret sharing scheme [52].
Generating this shared secret key either requires a trusted third party or a protocol for distributed key generation [30, 47]. The former is undesirable for most practical applications since it creates a single point of failure.
In a distributed key generation protocol, each peer contributes a number of preshares. The peers agree on the set of preshares and each peer recombines them in a different way, yielding the shares of the private threshold key.
In the key generation protocol used for the Cramer et al. voting scheme, the number of preshares that need to be agreed upon is quadratic in the number of peers. Every peer needs to know every preshare, even if it is not required by the individual peer for reconstructing the share, since the preshares are accompanied by noninteractive proofs of correctness. Thus, the number of values that need to be agreed upon is quadratic in the number of peers, which makes the use of setunion consensus attractive compared to individual agreement.
Even though the preshares can be checked for correctness, Byzantine consensus on the set of shares is still necessary for the case when a malicious peer submits a incorrect share to only some peers. Without Byzantine consensus, different correct recipients might exclude different peers, resulting in inconsistent shares.
Similarly, when a message that was encrypted with the threshold public key shall be decryped, every peer contributes a partial decryption with a proof of correctness. While the set of partial decryptions is typically linear in the number of peers, setunion consensus is still a reasonable choice here, this way the whole system only needs one agreement primitive.
7.3 Electronic voting with homomorphic encryption
Various conceptually different voting schemes use homomorphic encryption; we look as the scheme by Cramer et al. [17] as a modern and practical representative. A fundamental mechanism of the voting scheme is that a set of voting authorities A _{1},…,A _{ n } establish a threshold key pair that allows any entity that knows the public part of the key to encrypt a message that can only be decrypted when a threshold of the voting authorities cooperate. The homomorphism in the cryptosystem enables the computation of an encrypted tally with only the ciphertext of the submitted ballots. Ballots represent a choice of one candidate from a list of candidate options. The validity of encrypted ballot is ensured by equipping them with a noninteractive zeroknowledge proof of their validity.
It is assumed that the adversary is not able to corrupt more than 1/3 of the authorities. The voting process itself is then facilitated by all voters encrypting their vote and submitting it to the authorities. The encrypted tally is computed by every authority and then cooperatively decrypted by the authorities and published. Since correct authorities will only agree to decrypt the final tally and not individual ballots, the anonymity of the voter is preserved. For the voting scheme to work correctly, all correct peers must agree on exactly the same set of ballots before the cooperative decryption process starts; otherwise, the decryption of the tally will fail.
7.4 Other applications of BSC
Bitcoin [44] has gained immense popularity over the past few years. Bitcoin solves a slight variation of Byzantine consensus without strong validity [31, 39]. Given that a block in Bitcoin is basically just a set of (valid) transactions, BSC could be used to efficiently achieve agreement between participants about the next transaction group. Here, the most natural application would be to use BSC to improve the efficiency of proofofstake incentivized peers running BFT consensus in Cosmos [36].
8 Conclusions
Given m ballots, n authorities, f Byzantine faults and k ballots exclusively available to the adversary, voting with BSC achieves a complexity of O(m n+(f+k)n ^{3}), which in practice is better than the O(m n ^{2}) complexity of using SMR as m is usually significantly larger than n. Equivalent arguments hold for other applications requiring consensus over large sets. Furthermore, BSC remains advantageous in the absence of Byzantine failures, and the bounded set reconciliation makes it particularly efficient at handling various nonByzantine faults.
To ensure these performance bounds, BSC requires a bounded variant of Eppstein’s set reconciliation protocol that ensures that individual steps in the protocol cannot consume unbounded amounts of bandwidth. We are currently applying bounded set reconciliation in related domains, as any set reconciation can be made more robust if the complexity of the operation is bounded. For example, the GNU Name System [56] can use bounded set reconciliation when gossiping sets of key revocation sets. Here, the use of bounded set reconciliation protects the key revocation protocol against denialofservice attacks where an attacker might have previously sent excessively large IBFs or retransmitted known revocation messages already known to the recipient. The result is an efficient and resilient method for disseminating key revocation data.
In future work, it would be interesting to apply bounded set reconciliation to Byzantine consensus protocols that are more efficient than the simple gradecast consensus. It would also be interesting to experimentally compare bulletin boards using BSC with those using traditional replicated state machines.
9 Endnotes
^{1} Different variations and names can be found in the literature. We have chosen a definition that extends to our generalization to sets later on.
^{2} In practice, exponential backoff is used.
^{3} The ⊕ denotes bitwise exclusive or.
^{4} The optimal formula here depends on the size ratio of IBF element to the transmission size of an individual element and the estimated size of the set overlap. However, to simplify the exposition, we will assume a simple 50% threshold henceforth.
^{5} Each failure causes the IBF size to double and thus corresponds to a doubling of the set difference estimate. Thus, the number of decoding failures could remain the threshold that causes an abort, while the set difference estimate substantially exceeds \(2 (S_{i}  \mathcal {L})\).
^{6} An attempt has been made to relax the definition of SMC to alleviate this requirement, resulting in a weaker definition that includes nonunanimous aborts as a possible result [32]. This definition is mainly useful in scenarios without an nonfaulty 2/3 majority, where Byzantine consensus is not possible in the asynchronous model [22].
Notes
Acknowledgements
We thank Jeffrey Burdges and the anonymous reviewers of ARES 2016 for comments on an earlier draft of this paper.
Funding
This work benefits from the financial support of the Brittany Region (ARED 9174) and the Renewable Freedom Foundation.
Availability of data and materials
The software used for the experiments is available in the public Git repository [2].
Authors’ contributions
Both authors contributed to the design, implementation, and the writing of the article. Florian Dold ran the experiments. Both authors read and approved the final manuscript.
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
None.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
 1.The GNUnet Project. https://gnunet.org/. Accessed 28 Feb 2017.
 2.The GNUnet Project Git Repository. git://gnunet.org/git/gnunet. Accessed 28 Feb 2017.
 3.M AbdElMalek, GR Ganger, GR Goodson, MK Reiter, JJ Wylie, Faultscalable byzantine faulttolerant services. ACM SIGOPS Oper. Syst. Rev. 39(5), 59–74 (2005).CrossRefGoogle Scholar
 4.B Adida, in Proceedings of the 17th Conference on Security Symposium, SS’08. Helios: Webbased Openaudit Voting (USENIX AssociationBerkeley, 2008), pp. 335–348. http://dl.acm.org/citation.cfm?id=1496711.1496734.
 5.MK Aguilera, in Stumbling over Consensus Research: Misunderstandings and Issues, ed. by B CharronBost, F Pedone, and A Schiper. Replication (SpringerVerlagBerlin, 2010), pp. 59–72. http://dl.acm.org/citation.cfm?id=2172338.2172342.
 6.J Aspnes, Lower bounds for distributed coinflipping and randomized consensus. J. ACM (JACM). 45(3), 415–450 (1998).MathSciNetCrossRefMATHGoogle Scholar
 7.C Attiya, D Dolev, J Gil, in Proceedings of the Third Annual ACM Symposium on Principles of Distributed Computing, PODC ’84. Asynchronous Byzantine Consensus (ACMNew York, 1984), pp. 119–133. http://doi.acm.org/10.1145/800222.806740.
 8.PL Aublin, R Guerraoui, N Knežević, V Quéma, M Vukolić, The next 700 bft protocols. ACM Trans. Comput. Syst. 12(4), 1–12:45 (2015). doi: http://dx.doi.org/10.1145/2658994. http://doi.acm.org/10.1145/2658994.
 9.M BenOr, D Dolev, EN Hoch, Simple gradecast based algorithms (2010). arXiv preprint arXiv:1007.1049.Google Scholar
 10.JDC Benaloh, Verifiable secretballot elections (ACM, New Haven, 1987).MATHGoogle Scholar
 11.BH Bloom, Space/time tradeoffs in hash coding with allowable errors. Commun. ACM. 13(7), 422–426 (1970).CrossRefMATHGoogle Scholar
 12.P Bogetoft, DL Christensen, I Damgård, M Geisler, T Jakobsen, M Krøigaard, JD Nielsen, JB Nielsen, K Nielsen, J Pagter, M Schwartzbach, T Toft, in Financial Cryptography and Data Security, ed. by R Dingledine, P Golle. Secure multiparty computation goes live (SpringerVerlagBerlin, 2009), pp. 325–343. http://dx.doi.org/10.1007/9783642035494_20.
 13.C Cachin, K Kursawe, V Shoup, Random oracles in constantinople: practical asynchronous byzantine agreement using cryptography. J. Cryptol. 18(3), 219–246 (2005).MathSciNetCrossRefMATHGoogle Scholar
 14.M Castro, B Liskov, Practical byzantine fault tolerance and proactive recovery. ACM Trans. Comput. Syst. (TOCS). 20(4), 398–461 (2002).CrossRefGoogle Scholar
 15.M Castro, B Liskov, in Third Symposium on Operating Systems Design and Implementation (OSDI), 99. Practical Byzantine Fault Tolerance (USENIX Association, Cosponsored by IEEE TCOS and ACM SIGOPSNew Orleans, 1999), pp. 173–186.Google Scholar
 16.A Clement, E Wong, L Alvisi, M Dahlin, M Marchetti, in Proceedings of the 6th USENIX Symposium on Networked Systems Design and Implementation, NSDI’09. Making Byzantine Fault Tolerant Systems Tolerate Byzantine Faults (USENIX AssociationBerkeley, 2009), pp. 153–168. http://dl.acm.org/citation.cfm?id=1558977.1558988.
 17.R Cramer, R Gennaro, B Schoenmakers, A secure and optimally efficient multiauthority election scheme. Eur. Trans. Telecommun. 8(5), 481–490 (1997).CrossRefGoogle Scholar
 18.R De Prisco, D Malkhi, M Reiter, On kset consensus problems in asynchronous systems. IEEE Trans. Parallel Distrib. Syst. 12(1), 7–21 (2001).CrossRefMATHGoogle Scholar
 19.YG Desmedt, Threshold cryptography. Eur. Trans. Telecommun. 5(4), 449–458 (1994).CrossRefGoogle Scholar
 20.F Dold, Cryptographically secure, distributed electronic voting (2014). Bachelor’s thesis, Technische Universität München.Google Scholar
 21.D Dolev, C Dwork, L Stockmeyer, On the minimal synchronism needed for distributed consensus. J. ACM (JACM). 34(1), 77–97 (1987).MathSciNetCrossRefMATHGoogle Scholar
 22.C Dwork, N Lynch, L Stockmeyer, Consensus in the presence of partial synchrony. J. ACM (JACM). 35(2), 288–323 (1988).MathSciNetCrossRefGoogle Scholar
 23.D Eppstein, MT Goodrich, F Uyeda, G Varghese, What’s the Difference?: Efficient Set Reconciliation Without Prior Context. SIGCOMM Comput. Commun. Rev. 41(4), 218–229 (2011). http://doi.acm.org/10.1145/2043164.2018462.
 24.P Feldman, S Micali, in Proceedings of the Twentieth Annual ACM Symposium on Theory of Computing, STOC ’88. Optimal Algorithms for Byzantine Agreement (ACMNew York, 1988), pp. 148–161. http://doi.acm.org/10.1145/62212.62225.
 25.PN Feldman, Optimal algorithms for byzantine agreement (1988). Ph.D. thesis, Massachusetts Institute of Technology.Google Scholar
 26.MJ Fischer, NA Lynch, A lower bound for the time to assure interactive consistency (1981). Tech. rep. DTIC Document.Google Scholar
 27.MJ Fischer, NA Lynch, M Merritt, Easy impossibility proofs for distributed consensus problems. Distrib. Comput. 1(1), 26–39 (1986).CrossRefMATHGoogle Scholar
 28.MJ Fischer, NA Lynch, MS Paterson, Impossibility of distributed consensus with one faulty process. J. ACM (JACM). 32(2), 374–382 (1985).MathSciNetCrossRefMATHGoogle Scholar
 29.M Fitzi, M Hirt, in Proceedings of the Twentyfifth Annual ACM Symposium on Principles of Distributed Computing, PODC ’06. Optimally Efficient Multivalued Byzantine Agreement (ACMNew Yorks, 2006), pp. 163–168. http://doi.acm.org/10.1145/1146381.1146407.
 30.PA Fouque, J Stern, in Public Key Cryptography: 4th International Workshop on Practice and Theory in Public Key Cryptosystems, PKC 2001 Cheju Island, Korea, February 13–15, 2001 Proceedings, ed. by K Kim. One Round Threshold DiscreteLog Key Generation without Private Channels (SpringerBerlin, 2001), pp. 300–316. https://doi.org/10.1007/3540445862_22.
 31.J Garay, A Kiayias, N Leonardos, M Fischlin, in Advances in Cryptology  EUROCRYPT 2015: 34th Annual International Conference on the Theory and Applications of Cryptographic Techniques, Sofia, Bulgaria, April 2630, 2015, Proceedings, Part II, ed. by E Oswald. The Bitcoin Backbone Protocol: Analysis and Applications (SpringerBerlin, 2015), pp. 281–310. https://doi.org/10.1007/9783662468036_10.
 32.S Goldwasser, Y Lindell, Secure multiparty computation without agreement. J. Cryptol. 18(3), 247–287 (2005).MathSciNetCrossRefMATHGoogle Scholar
 33.R Guerraoui, M Hurfinn, A Mostefaoui, R Oliveira, M Raynal, A Schiper, in Advances in Distributed Systems: Advanced Distributed Computing: From Algorithms to Systems, ed. by S Krakowiak, S Shrivastava. Consensus in Asynchronous Distributed Systems: A Concise Guided Tour (SpringerBerlin, 2000), pp. 33–47. https://doi.org/10.1007/3540464751_2.
 34.KP Kihlstrom, LE Moser, PM MelliarSmith, in Proceedings of the ThirtyFirst Annual Hawaii International Conference on System Sciences  Volume 3, HICSS ’98. The SecureRing Protocols for Securing Group Communication (IEEE Computer SocietyWashington, 1998), pp. 317–326. http://dx.doi.org/10.1109/HICSS.1998.656294.
 35.R Kotla, L Alvisi, M Dahlin, A Clement, E Wong, in Proceedings of Twentyfirst ACM SIGOPS Symposium on Operating Systems Principles, SOSP ’07. Zyzzyva: Speculative Byzantine Fault Tolerance (ACMNew York, 2007), pp. 45–58. http://doi.acm.org/10.1145/1294261.1294267.
 36.J Kwon, E Buchman, Cosmos: a network of distributed ledgers (2016). https://cosmos.network/whitepaper. Accessed 22 Feb 2017.
 37.L Lamport, R Shostak, M Pease, The byzantine generals problem. ACM Trans. Program. Lang. Syst. (TOPLAS). 4(3), 382–401 (1982).CrossRefMATHGoogle Scholar
 38.N Malpani, JL Welch, N Vaidya, in Proceedings of the 4th international workshop on Discrete algorithms and methods for mobile computing and communications. Leader election algorithms for mobile ad hoc networks (ACMNew York, 2000), pp. 96–103.Google Scholar
 39.A Miller, JJ LaViola Jr, Anonymous byzantine consensus from moderatelyhard puzzles: A model for bitcoin, CSTR1401 (University of Central Florida, 2014). http://decentralize.ece.illinois.edu/publications/.
 40.A Miller, Y Xia, K Croman, E Shi, D Song, in Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, CCS ’16. The honey badger of bft protocols (ACMNew York, 2016), pp. 31–42. doi: http://dx.doi.org/10.1145/2976749.2978399. http://doi.acm.org/10.1145/2976749.2978399.
 41.Y Minsky, A Trachtenberg, R Zippel, Set reconciliation with nearly optimal communication complexity. IEEE Trans. Inf. Theory. 49(9), 2213–2218 (2003).MathSciNetCrossRefMATHGoogle Scholar
 42.M Mitzenmacher, R Pagh, Simple multiparty set reconciliation (2013). arXiv preprint arXiv:1311.2037.Google Scholar
 43.A Mostefaoui, H Moumen, M Raynal, in Proceedings of the 2014 ACM symposium on Principles of distributed computing. Signaturefree asynchronous byzantine consensus with t< n/3 and o (n 2) messages (ACMNew York, 2014), pp. 2–9.Google Scholar
 44.S Nakamoto, Bitcoin: A peertopeer electronic cash system. Consulted. 1:, 28 (2012).Google Scholar
 45.G Neiger, Distributed consensus revisited. Inf. Process. Lett. 49(4), 195–201 (1994).CrossRefMATHGoogle Scholar
 46.C Okasaki, Purely functional data structures (Cambridge University Press, New York, 1999).MATHGoogle Scholar
 47.TP Pedersen, in Advances in Cryptology—EUROCRYPT’91. A threshold cryptosystem without a trusted party (Springer, 1991), pp. 522–526.Google Scholar
 48.R Peters, A secure bulletin board (2005). Master’s thesis, Technische Universiteit Eindhoven.Google Scholar
 49.B Polot, C Grothoff, in Ad Hoc Networking Workshop (MEDHOCNET), 2014 13th Annual Mediterranean. Cadet: Confidential adhoc decentralized endtoend transport (IEEE, 2014), pp. 71–78.Google Scholar
 50.MK Reiter, in Theory and Practice in Distributed Systems. The rampart toolkit for building highintegrity services (SpringerLondon, 1995), pp. 99–110.CrossRefGoogle Scholar
 51.J Saia, M Zamani, in SOFSEM 2015: Theory and Practice of Computer Science. Recent results in scalable multiparty computation (SpringerBerlin, 2015), pp. 24–44.Google Scholar
 52.A Shamir, How to share a secret. Commun. ACM. 22(11), 612–613 (1979).MathSciNetCrossRefMATHGoogle Scholar
 53.E Syta, P Jovanovic, EK Kogias, N Gailly, L Gasser, I Khoffi, MJ Fischer, B Ford, Scalable biasresistant distributed randomness (2016). Cryptology ePrint Archive, Report 2016/1067. http://eprint.iacr.org/2016/1067. Accessed 22 Feb 2017.
 54.S Tarkoma, CE Rothenberg, E Lagerspetz, Theory and practice of bloom filters for distributed systems. IEEE Commun. Surv. Tutor. 14(1), 131–155 (2012).CrossRefGoogle Scholar
 55.SH Totakura, Large scale distributed evaluation of peertopeer protocols (2013). Master’s thesis, Technische Universität München, Garching bei München.Google Scholar
 56.M Wachs, M Schanzenbach, C Grothoff, in Proceedings of the 13th International Conference on Cryptology and Network Security  Volume 8813. A CensorshipResistant, PrivacyEnhancing and Fully Decentralized Name System (SpringerVerlag Inc.New York, 2014), pp. 127–142. http://dx.doi.org/10.1007/9783319122809_9.
Copyright information
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.