1 Introduction

Block ciphers are one of the most important building blocks in many security protocols. Modern block ciphers are cascades of several rounds where every round consists of confusion and diffusion layers. In many block ciphers, while the confusion layer is often realized as a parallel application of non-linear substitution boxes (S-boxes), the diffusion layer is built from a linear transformation. The diffusion layer plays an efficacious role in providing resistance against the most well-known attacks on block ciphers, such as differential cryptanalysis (DC) [2], and linear cryptanalysis (LC) [8].

When considering a word-based linear transformation, where the word size is equal to the input/output size of the S-box, the branch number provides a lower bound on the number of active S-boxes throughout the diffusion layer for differential and linear attacks. The goal for a designer is to maximize this number, in order to diffuse the non-linear properties of the S-Boxes faster to the subsequent rounds of the cipher. The faster this non-linearity spreads, the less number of rounds the cipher requires to become secure against linear and differential attacks. It has been shown that the maximal branch number for a linear transformation of s words is s+1 and diffusion layers with maximal branch number can be achieved by using MDS matrices [4].

An MDS matrix (Maximum Distance Separable) is a matrix representing a function with certain diffusion properties that have useful applications in cryptography. Technically, an m×n matrix A over a finite field K is an MDS matrix if it is the transformation matrix of a linear transformation f(x)=Ax from K n to K m such that no two different (m+n)-tuples of the form (x,f(x)) coincide in n or more components. Equivalently, the set of all (m+n)-tuples (x,f(x)) is an MDS code, i.e. a linear code that reaches the Singleton bound.

In 1994, Vaudenay [11, 12] suggested using MDS matrices in cryptographic primitives to produce what he called multipermutations, not-necessarily linear functions with the same property. These functions have what he called perfect diffusion: changing t of the inputs change at least mt+1 of the outputs. He showed how to exploit imperfect diffusion to cryptanalyze functions that are not multipermutations. MDS matrices were later used in many block ciphers such as Square, SHARK, AES, Twofish and Hierocrypt and in the stream cipher MUGI and the cryptographic hash function Whirlpool.

The common approach to construct MDS matrices is to extract them from MDS codes such as Reed–Solomon codes [7]. However, constructing MDS diffusion layers with low-cost implementations is a challenge for designers. Another problem arises when MDS diffusion layers are exploited in substitution-permutation networks (SPN), where the MDS matrix is used in the encryption and its inverse is used in the decryption process. Thus, constructing MDS matrices with low-cost inverse is of great importance.

In this paper, we propose a new method to construct low-cost diffusion layers with an extra property that their inverse can also be implemented efficiently. We call the proposed layer a recursive diffusion layer. It is constructed from several rounds of Feistel-like structures whose round functions are linear. It consists of simple linear operations such as shift, rotation and XOR with very similar inversion operations. We are going to elaborate on the conditions for the underlying linear function to be an MDS matrix using one or multiple such linear functions by proposing a systematic method to find them. We believe that our proposed solution would be a rather simple recipe for designing a diffusion layer with maximal branch number and will be useful for future designs of cryptographic algorithms.

1.1 Notations

Let x be an array of s n-bit elements x=[x 0(n),x 1(n),…,x s−1(n)]. The number of non-zero elements in x is denoted by w(x), also known as the Hamming weight of x. The following notations are used throughout this paper:

figure a

For a diffusion layer D applicable on x, we have the following definitions:

Definition 1

([4])

The differential branch number of a linear diffusion layer D is defined as

$$\beta_{d}(D)=\displaystyle\min_{\mathbf{x}\not=0}\bigl\{ w(\mathbf {x})+w\bigl(D(\mathbf{x})\bigr)\bigr\} $$

We know that the linear function D can be shown as a binary matrix B, and D t is a linear function obtained from B t, where B t is the transposition of B.

Definition 2

([4])

The linear branch number of a linear diffusion layer D is defined as:

$$\beta_{l}(D)=\displaystyle\min_{\mathbf{x}\not=0}\bigl\{ w(\mathbf {x})+w\bigl(D^{t}(\mathbf{x})\bigr)\bigr\} $$

It is well known that for a diffusion layer acting on s-word inputs, the maximal β d and β l are s+1 [4]. A diffusion layer D taking its maximal β d and β l is called a perfect or MDS diffusion layer. Furthermore, a diffusion layer with β d =β l =s is called an almost perfect diffusion layer [4].

1.2 Our Contribution

In this paper, we define the notion of a recursive diffusion layer, and we propose a method to construct such perfect diffusion layers.

Definition 3

A diffusion layer D with s words x i as the input and s words y i as the output is called a recursive diffusion layer if it can be represented in the following form:

$$ D: \left\{ \begin{array}{l} y_{0}= x_{0} \oplus F_0(x_{1}, x_{2},\ldots, x_{s-1})\\ y_{1}= x_{1} \oplus F_1(x_{2}, x_{3},\ldots, x_{s-1}, y_0)\\ \vdots\\ y_{s-1}= x_{s-1} \oplus F_{s-1}(y_{0}, y_{1},\ldots, y_{s-2}) \end{array} \right. $$
(1)

where F 0,F 1,…,F s−1 are arbitrary linear functions.

An advantage of this structure is that the inverse of D is very similar to D and does not require the inverse of F i functions. The inverse can be computed as:

$$D^{-1}: \left\{ \begin{array}{l} x_{s-1}= y_{s-1} \oplus F_{s-1}(y_{0}, y_{1},\ldots, y_{s-2})\\ x_{s-2}= y_{s-2} \oplus F_{s-2}(x_{s-1}, y_{0},\ldots, y_{s-3})\\ \vdots\\ x_{0}= y_{0} \oplus F_{0}(x_{1}, x_{2},\ldots, x_{s-1}) \end{array} \right. $$
(2)

As an example, consider a 2-round Feistel structure with a linear round function L as a recursive diffusion layer with s=2. The input–output relation for this diffusion layer is

$$ D: \left\{ \begin{array}{l} y_{0}= x_{0} \oplus L(x_{1})\\ y_{1}= x_{1} \oplus L( y_0) \end{array} \right. $$

The quarter-round function of the stream cipher Salsa20 is an example of a non-linear recursive diffusion layer [1].

$$ D: \left\{ \begin{array}{l} y_{1}= x_{1} \oplus((x_0+x_3) \lll7)\\ y_{2}= x_{2} \oplus((x_0+y_1) \lll9)\\ y_{3}= x_{3} \oplus((y_1+y_2)\lll13)\\ y_{0}= x_{0} \oplus((y_2+y_3) \lll18)\\ \end{array} \right. $$

Also, the lightweight hash function PHOTON [5] and the block cipher LED [6] use MDS matrices based on Eq. (1). In these ciphers, an m×m MDS matrix B m was designed based on the following matrix B for the performance purposes:

$$\mathbf{B} = \left(\begin{array}{c@{\quad }c@{\quad }c@{\quad }c@{\quad }c} 0 & 1 & 0 & \ldots& 0\\ 0& 0 & 1 & \ldots& 0\\ \vdots& & \ddots& \\ 0& 0 & 0 & \ldots& 1\\ 1 & Z_{1} & Z_{2} & \ldots& Z_{m-1} \end{array} \right) $$

By matrix B, one element of m inputs is updated and other elements are shifted. If we use B m, all inputs are updated, but we must check if this matrix is MDS. One example for m=4 is the PHOTON matrix working over GF(28):

$$ \mathbf{B} = \left(\begin{array}{c@{\quad}c@{\quad}c@{\quad}c} 0 & 1 & 0 & 0\\ 0& 0 & 1 & 0\\ 0& 0 & 0 & 1\\ 1& 2 & 1 & 4\\ \end{array}\right) \quad \Rightarrow \quad \mathbf{B}^4 = \left(\begin{array}{c@{\quad}c@{\quad}c@{\quad}c} 1 & 2 & 1 & 4\\ 4& 9 & 6 & 17\\ 17& 38 & 24 & 66\\ 66& 149 & 100 & 11\\ \end{array}\right) $$

In this paper, we propose a new approach to design linear recursive diffusion layers with the maximal branch number in which F i ’s are composed of one or two linear functions and a number of XOR operations. The design of the proposed diffusion layer is based on the invertibility of some simple linear functions in GF(2). Linear functions in this diffusion layer can be designed to be low-cost for different sizes of the input words, thus the proposed diffusion layer might be appropriate for resource-constrained devices, such as RFID tags. Although these recursive diffusion layers are not involutory, they have similar inverses with the same computational complexity.

This paper proceeds as follows: In Sect. 2, we introduce the general structure of our proposed recursive diffusion layer. Then, for one of its instances, we systematically investigate the required conditions for the underlying linear function to achieve the maximal branch number. In Sect. 3, we propose some other recursive diffusion layers with less than 8 input words and only one linear function. We use two linear functions to have a perfect recursive diffusion layer for s>4 in Sect. 4. Finally, we conclude the paper in Sect. 5.

2 The Proposed Diffusion Layer

In this section, we introduce a new perfect linear diffusion layer with a recursive structure. The diffusion layer D takes s words x i for i={0,1,…,s−1} as input, and returns s words y i for i={0,1,…,s−1} as output. So, we can represent this diffusion layer as

$$y_{0}||y_{1}||\cdots||y_{s-1}=D(x_{0}||x_{1}|| \cdots||x_{s-1}) $$

The first class of the proposed diffusion layer D is represented in Fig. 1, where L is a linear function, α k ,β k ∈{0,1}, α 0=1 and β 0=0. This diffusion layer can be represented in the form of Eq. (1) in which the F i functions are all the same and can be represented as

$$F_i(x_1,x_2,\ldots,x_{s-1})= \bigoplus_{j=1}^{s-1} \alpha_{j}x_j \oplus L \Biggl(\bigoplus_{j=1}^{s-1} \beta_{j}x_j \Biggr) $$
Fig. 1.
figure 1

The first class of the recursive diffusion layers.

To guarantee the maximal branch number for D, the linear function L and the coefficients α j and β j must satisfy some necessary conditions. Conditions on L are expressed in this section and those of α j ’s and β j ’s are expressed in Sect. 3. The diffusion layer described by Eq. (3) is an instance that satisfies the necessary conditions on α j , and β j with s=4. In the rest of this section, we concentrate on the diffusion layers of this form and show that we can find invertible linear functions L such that D becomes a perfect diffusion layer.

$$ D: \left\{ \begin{array}{l} y_{0}= x_{0} \oplus x_2 \oplus x_3 \oplus L(x_1 \oplus x_3)\\ y_{1}= x_{1} \oplus x_3 \oplus y_0 \oplus L(x_2 \oplus y_0)\\ y_{2}= x_{2} \oplus y_0 \oplus y_1 \oplus L(x_3 \oplus y_1)\\ y_{3}= x_{3} \oplus y_1 \oplus y_2 \oplus L(y_0 \oplus y_2) \end{array} \right. $$
(3)

As shown in Fig. 2, this diffusion layer has a Feistel-like (GFN) structure, i.e.,

$$F_0(x_1,x_2,x_3)=x_2 \oplus x_3 \oplus L(x_1 \oplus x_3) $$
Fig. 2.
figure 2

The proposed recursive diffusion layer of Eq. (3).

The inverse transformation, D −1, has a very simple structure and does not require the inversion of the linear function L. The inverse of D is

$$ D^{-1}: \left\{ \begin{array}{l} x_{3}= y_{3} \oplus y_1 \oplus y_2 \oplus L(y_0 \oplus y_2)\\ x_{2}= y_{2} \oplus y_0 \oplus y_1 \oplus L(x_3 \oplus y_1)\\ x_{1}= y_{1} \oplus x_3 \oplus y_0 \oplus L(x_2 \oplus y_0)\\ x_{0}= y_{0} \oplus x_2 \oplus x_3 \oplus L(x_1 \oplus x_3) \end{array} \right. $$

D and D −1 are different, but they have the same structure and properties. To show that D has the maximal branch number, first we introduce some lemmas and theorems.

If L(x) can be written as ax in a finite field, then Eq. (3) can be expressed as a matrix representation as below:

$$ \mathbf{B} = \left(\begin{array}{c@{\quad}c@{\quad}c@{\quad}c} 0 & 1 & 0 & 0\\ 0 & 0 & 1 & 0\\ 0 & 0 & 0 & 1\\ 1 & a & 1 & a+1\\ \end{array}\right) \quad \Rightarrow \quad \left(\begin{array}{c} y_0 \\ y_1\\ y_2\\ y_3\\ \end{array}\right) =\mathbf{B}^4 \left(\begin{array}{c} x_0 \\ x_1\\ x_2\\ x_3\\ \end{array}\right) $$
(4)

We can construct MDS matrix similar to PHOTON matrix by the proposed diffusion layer. In Eq. (1), if F i (x 1,x 2,x 3)=F 0(x 1,x 2,x 3)=L(x 1)⊕x 2L 2(x 3), where L(x)=2x and xGF(28), PHOTON MDS matrix is obtained [5]. If we change B to Eq. (3), and define L(x)=2x, we have

$$ \mathbf{B} = \left(\begin{array}{c@{\quad}c@{\quad}c@{\quad}c} 0 & 1 & 0 & 0\\ 0& 0 & 1 & 0\\ 0& 0 & 0 & 1\\ 1& 2 & 1 & 3\\ \end{array}\right) \quad \Rightarrow \quad \mathbf{B}^4 = \left(\begin{array}{c@{\quad}c@{\quad}c@{\quad}c} 1 & 2 & 1 & 3\\ 3& 7 & 1 & 4\\ 4& 11 & 3 & 13\\ 13& 30 & 6 & 20\\ \end{array}\right) $$

Theorem 4

([4])

A Boolean function F has maximal differential branch number if, and only if it has maximal linear branch number.

As a result of Theorem 4, if we prove that the diffusion layer D represented in Eq. (3) has the maximal differential branch number, its linear branch number will be maximal too. Thus, in the following, we focus on the differential branch number.

Lemma 5

A linear functions L(x) is invertible if, and only if for any non-zero value a, \(L(a)\not=0\).

Proof

For any linear function L(x), we have L(0)=0. If there exists \(a\not =0\) such that L(a)=0, then L(x) is not invertible. On the other hand, suppose a=0 is the unique zero of L(x), and L(x) is not invertible. So, there exist two values b and c (\(b\not=c\)) such that L(b)=L(c). Since L(x) is a linear function, we have L(bc)=L(b)⊕L(c)=0, while \(b\oplus c\not=0\). This contradicts the assumption that a=0 is the unique zero of L(x). □

Lemma 6

Assume the linear operator \(\mathcal{L}_{i}\) corresponds to the linear function L i (x). If the linear operator \(\mathcal{L}_{3}\) can be represented as the multiplication of two operators \(\mathcal{L}_{1}\) and \(\mathcal{L}_{2}\), then the corresponding linear function L 3(x)=L 2(L 1(x)) is invertible if, and only if the linear functions L 1(x) and L 2(x) are invertible.

Proof

If L 1(x) and L 2(x) are invertible, clearly L 3(x) is invertible too. On the other hand, if L 3(x) is invertible then L 1(x) must be invertible, otherwise, there are distinct x 1, and x 2 such that L 1(x 1)=L 1(x 2). Thus, L 3(x 1)=L 2(L 1(x 1))=L 2(L 1(x 2))=L 3(x 2) which contradicts the invertibility of L 3(x). The invertibility of L 2(x) is proved in the same way. □

Example 1

We can rewrite the linear function L 3(x)=L 3(x)⊕x (\(\mathcal{L}_{3}=\mathcal{L}^{3} \oplus I\)) as L 3(x)=L 2(L 1(x)), where L 1(x)=L(x)⊕x (\(\mathcal {L}_{1}=\mathcal{L} \oplus I\)) and L 2(x)=L 2(x)⊕L(x)⊕x (\(\mathcal{L}_{2}=\mathcal{L}^{2} \oplus \mathcal{L} \oplus I\)). Thus, the invertibility of L 3(x) is equivalent to the invertibility of the two linear functions L 1(x) and L 2(x).

Theorem 7

For the diffusion layer represented in Eq. (3), if the four linear functions L(x), xL(x), xL 3(x) and xL 7(x) are invertible, then this diffusion layer is perfect.

Proof

We show that the differential branch number of this diffusion layer is 5. First, the 4 words of the output are directly represented as functions of the 4 words of the input:

$$ D: \left\{ \begin{array}{l} y_{0}= x_{0} \oplus L(x_1 ) \oplus x_2 \oplus x_3 \oplus L(x_3)\\ y_{1}= x_0 \oplus L(x_0) \oplus x_1 \oplus L(x_1) \oplus L^2(x_1) \oplus x_2 \oplus L^2(x_3)\\ y_{2}= L^2(x_0) \oplus x_1 \oplus L(x_1) \oplus L^3(x_1) \oplus x_2 \oplus L(x_2)\oplus x_3 \oplus L^2(x_3) \oplus L^3(x_3)\\ y_{3}= x_0 \oplus L^2(x_0) \oplus L^3(x_0) \oplus L(x_1) \oplus L^2(x_1) \oplus L^3(x_1) \oplus L^4(x_1) \\ \phantom{y_{3}=} \oplus L(x_2) \oplus L^2(x_2) \oplus L^2(x_3) \oplus L^4(x_3) \end{array} \right. $$
(5)

In the proof, we look at all different cases for the Hamming weight of the input. In other words, we show that if the Hamming weight of the input is m=1,2,3,4, then the Hamming weight of the output is greater than or equal to 5−m. Each case will pose different conditions on L which in the end can be summarized to the condition given in the theorem. The diffusion layer represented in Eq. (3) is invertible. Consider m=4, then all of the 4 words in the input are active, and we are sure at least one of the output words is active too. Thus, the theorem is correct for m=4. The remainder of the proof is performed for the 3 cases of w(Δ(x))=m, for m=1,2,3 separately. In each of these cases, some conditions are forced on the linear function L.

Case 1: w(△x)=1

To study this case, first the subcase

$$(\triangle x_0\not=0,\triangle x_{1}=\triangle x_{2}=\triangle x_{3}=0 \quad \mbox{or}\quad \triangle \mathbf{x}=\triangle x_0|| 0|| 0||0 ) $$

is analyzed. For this subcase, Eq. (5) is simplified to:

$$ D: \left\{ \begin{array}{l} \triangle y_{0}= \triangle x_{0}\\ \triangle y_{1}=(I\oplus L) (\triangle x_0) \\ \triangle y_{2}= L^2(\triangle x_0)\\ \triangle y_{3}= (I\oplus L^2 \oplus L^3)(\triangle x_0 ) \end{array} \right. $$

If D is a perfect diffusion layer, then △y 0, △y 1, △y 2, and △y 3 must be non-zero. Clearly, △y 0 is non-zero and based on Lemma 5, the conditions for △y 1, △y 2, and △y 3 to be non-zero are that the linear functions IL, L 2, and IL 2L 3 must be invertible. Note that based on Lemma 6 the invertibility of L yields the invertibility of L 2. Considering Lemma 6, if the other three sub-cases are studied, it is induced that the linear functions xL(x)⊕L 2(x) and xL(x)⊕L 3(x) must also be invertible.

Case 2: w(△x)=2

In this case, there exist exactly two active words in the input difference, and we obtain some conditions on the linear function L to guarantee the branch number 5 for D. In the following, we only analyze the subcase

$$(\triangle x_0,\triangle x_{1}\not=0 \quad\mbox{and}\quad \triangle x_{2}=\triangle x_{3}=0 \quad\mbox{or}\quad \triangle\mathbf{x}=\triangle x_0|| \triangle x_1|| 0||0 ) $$

With this assumption, Eq. (5) is simplified to

$$ D: \left\{ \begin{array}{l} \triangle y_{0}= \triangle x_{0} \oplus L(\triangle x_1 ) \\ \triangle y_{1}=(I\oplus L) (\triangle x_0) \oplus(I\oplus L \oplus L^2)( \triangle x_1) \\ \triangle y_{2}= L^2(\triangle x_0) \oplus(I\oplus L \oplus L^3)(\triangle x_1) \\ \triangle y_{3}= (I\oplus L^2 \oplus L^3)(\triangle x_0 ) \oplus (L\oplus L^2\oplus L^3\oplus L^4)(\triangle x_1) \end{array} \right. $$
(6)

To show that w(△y) is greater than or equal to 3, we must find some conditions on L such that if one of the △y i ’s is zero, then the other three △y j ’s cannot be zero. Let △y 0=0, then

$$ \triangle x_{0} \oplus L(\triangle x_1)=0\quad \Rightarrow\quad \triangle x_{0} = L(\triangle x_1) $$

If △x 0 is replaced in the last three equations of Eq. (6), we obtain △y 1, △y 2 and △y 3 as follows:

$$ \left\{ \begin{array}{l} \triangle y_{1}=\triangle x_1\\ \triangle y_{2}=\triangle x_1 \oplus L(\triangle x_1)\\ \triangle y_{3}= L^2(\triangle x_1)\\ \end{array} \right. $$

Obviously, △y 1 is not zero. Furthermore, considering Lemma 5, for △y 2 to be non-zero, we conclude that the function xL(x) must be invertible. For △y 1⇒△y 3, L 2(x) is invertible. This condition was already obtained in the Case 1. We continue this procedure for △y 1=0.

$$ \begin{array}{l} \triangle y_{1}= \triangle x_0 \oplus L(\triangle x_0) \oplus x_1 \oplus L(\triangle x_1) \oplus L^2(\triangle x_1)=0\\ \quad\Rightarrow\quad \triangle x_0 \oplus L(\triangle x_0) = x_1 \oplus L(\triangle x_1) \oplus L^2(\triangle x_1) \end{array} $$

From the previous subcase, we know that if △y 0=0, then \(\triangle y_{1}\not=0\). Thus, we conclude that △y 0 and △y 1 cannot be simultaneously zero. Therefore, by contraposition, we obtain that if △y 1=0, then \(\triangle y_{0}\not=0\). So, we only check △y 2 and △y 3. From the third equation in Eq. (6), we have

$$\begin{array}{lll} (I \oplus L) (\triangle y_{2})&=&L^2(\triangle x_1) \oplus L^3(\triangle x_1) \oplus L^4(\triangle x_1) \oplus\triangle x_1\\ &&\oplus\, L^2(\triangle x_1) \oplus L^3(\triangle x_1) \oplus L^4(\triangle x_1)\\ &=&\triangle x_1 \end{array} $$

xL(x) is invertible, thus we conclude that with the two active words △x 0 and △x 1 in the input, △y 1 and △y 2 cannot be zero simultaneously. With the same procedure, we can prove that △y 1, and △y 3 cannot be zero simultaneously.

Here we only gave the proof for the case (\(\triangle x_{0},\triangle x_{1}\not=0\), △x 2=△x 3=0). We performed the proof procedure for the other cases, and no new condition was added to the previous set of conditions in Case 1.

Case 3: w(△x)=3

In this case, assuming three active words in the input, we show that the output has at least 2 non-zero words. Here, only the case

$$(\triangle x_{0},\triangle x_{1},\triangle x_{2}\not=0 \quad\mbox{and}\quad \triangle x_{3}=0 \quad \mbox{or}\quad \triangle\mathbf{x} = \triangle x_0 || \triangle x_1|| \triangle x_{2} || 0 ) $$

is analyzed. The result holds for the other three cases with w(△x)=3. Let rewrite Eq. (5) for △x 3=0 as follows:

$$ D: \left\{ \begin{array}{l} \triangle y_{0}= \triangle x_{0} \oplus L(\triangle x_1 ) \oplus \triangle x_2 \\ \triangle y_{1}=(I\oplus L) (\triangle x_0) \oplus(I\oplus L \oplus L^2)( \triangle x_1) \oplus\triangle x_2 \\ \triangle y_{2}= L^2(\triangle x_0) \oplus(I\oplus L \oplus L^3)(\triangle x_1) \oplus(I \oplus L) (\triangle x_2 )\\ \triangle y_{3}= (I\oplus L^2 \oplus L^3)(\triangle x_0 ) \oplus (L\oplus L^2\oplus L^3\oplus L^4)(\triangle x_1) \oplus (L\oplus L^2)(\triangle x_2) \end{array} \right. $$
(7)

When △y 0=△y 1=0, from the first two lines of Eq. (7), △x 0 and △x 1 are obtained as the function of △x 2.

$$ \left\{ \begin{array}{lll} \triangle y_{0}&=& \triangle x_{0} \oplus L(\triangle x_1 ) \oplus \triangle x_2 =0\\ \triangle y_{1}&=& \triangle x_0 \oplus L(\triangle x_0) \oplus \triangle x_1 \oplus L(\triangle x_1)\\ &&\oplus L^2(\triangle x_1) \oplus \triangle x_2=0 \end{array} \right. \quad \Rightarrow \quad \left\{ \begin{array}{l} \triangle x_{1} =L( \triangle x_2)\\ \triangle x_{0}= \triangle x_2 \oplus L^2(\triangle x_2) \end{array} \right. $$

Now, replacing △x 0=△x 2L 2(△x 2) and △x 1=L(△x 2) into △y 2 and △y 3 yields

$$ \left\{ \begin{array}{rcl} \triangle y_{2}&=& L^2(\triangle x_0) \oplus(I\oplus L \oplus L^3)(\triangle x_1) \oplus(I \oplus L) (\triangle x_2 )=\triangle x_2\\ \triangle y_{3}&=& (I\oplus L^2 \oplus L^3)(\triangle x_0 ) \oplus (L\oplus L^2\oplus L^3\oplus L^4)(\triangle x_1) \oplus (L\oplus L^2)(\triangle x_2)\\ &=&(I\oplus L)(\triangle x_2) \end{array} \right. $$

From Case 1, we know that the functions xL(x) are invertible. Therefore, △y 2, and △y 3 are non-zero. If the other sub-cases with three active words in the input are investigated, it is easy to see that no new condition is added to the present conditions on L. Finally, we conclude that the diffusion layer D presented in Fig. 1 is perfect if the linear functions

$$ \left\{ \begin{array}{l} L_1(x)=L(x)\\ L_2(x)=x \oplus L(x)\\ L_3(x)=x \oplus L(x) \oplus L^2(x)\\ L_4(x)=x \oplus L(x) \oplus L^3(x)\\ L_5(x)=x \oplus L^2(x) \oplus L^3(x) \end{array} \right. $$

are invertible. We know that L 3(L 2(x))=xL 3(x) and L 5(L 4(L 2(x)))=xL 7(x). Thus, by Lemma 6, we can summarize the necessary conditions on the linear function L as the invertibility of L(x), (IL)(x), (IL 3)(x), and (IL 7)(x). □

Next, we need a simple method to check whether a linear function L satisfies the conditions of Theorem 7 or not. For this purpose, we use the binary matrix representation of L. Assume that x i is an n-bit word. Hence, we can represent a linear function L with an n×n matrix A with elements in GF(2). As a result of Lemma 5, if L is invertible, A is not singular over GF(2) (\(|\mathbf{A}|\not=0\)). To investigate whether a linear function L satisfies the conditions of Theorem 7, we construct the corresponding matrix A n×n from L, and check the non-singularity of the matrices A, IA, IA 3, and IA 7 in GF(2).

In the following, we construct concrete functions L which are lightweight and satisfy the conditions mentioned in Theorem 7. For example, the functions L(x)=x, L(x)=xa and L(x)=xa are the examples of the most lightweight linear functions. However, they do not satisfy Theorem 7 conditions, because at least one of the two functions L(x) and xL(x) are not invertible. A set of candidates for lightweight linear functions can be expressed as:

$$ \begin{array}{l} L(x_{(n)}) = (x_{(n)} \ll a) \oplus(x_{(n)} \gg b) \end{array} . $$
(8)

If (a+b)|n, then L(x) is invertible [15]. The remaining conditions xL(x), xL 3(x) and xL 7(x) have to be checked. Although the linear function in Eq. (8) has a complicated inverse, it does not require circular shift which is considered as an advantage for this function. Note that circular shift is not supported by some compilers. Another proposal for L(x) is

$$ \begin{array}{l} L(x_{(n)}) = \bigl(x_{(n)} \oplus(x_{(n)} \gg a)\bigr) \lll b \end{array} $$
(9)

The linear function in Eq. (9), for a>n/2, has a lightweight inverse L(x (n))=(x (n)b)⊕(x (n)b)≫a which will be used in diffusion layer proposed in Sect. 4.

We introduce some lightweight linear functions with n-bit inputs/outputs in Table 1 which satisfy the conditions of Theorem 7. Note that for n=8, there does not exist any linear function of the form Eq. (8) or Eq. (9) satisfying conditions of Theorem 7.

Table 1. Some instances of the linear function L satisfying Theorem 7.

2.1 Application of the Proposed Diffusion Layer in Current Block Ciphers

Together with designing new lightweight block ciphers, the proposed diffusion layer can also be applied to diffuse the non-linearity of big-size S-boxes. One of these block ciphers is MMB [3] that uses 32-bit S-boxes. Each round of MMB is composed of four transformations:

  • σ: bit-wise XOR of the intermediate value and the round key.

  • γ: modular multiplication of each 32-bit word of the intermediate value with a fixed 32-bit constant G i modulo 232−1.

  • η: an operation on two of the four input words.

  • θ: the only diffusion operation in MMB which is an involutory binary matrix as below:

    $$ \mathbf{B} = \left(\begin{array}{c@{\quad}c@{\quad}c@{\quad}c} 1 & 1 & 0 & 1\\ 1& 1 & 1 & 0\\ 0 & 1 & 1 & 1\\ 1& 0 & 1 & 1\\ \end{array}\right) $$

We can use the proposed diffusion layer with L(x (32))=(x (32)≪3)⊕(x (32)≫1) instead of the diffusion layer used in the block cipher MMB. If we use the proposed diffusion layer in this cipher, it becomes stronger against differential and linear attacks, because branch number of the binary matrix of MMB is 4 while branch number of the proposed diffusion layer is 5. This change also prevents the attacks presented against this block cipher in [13]. By computer simulations in C using a PC with CPU: 2.93 GHz and RAM: 2 GB, we observed that this modification reduces the performance of MMB by making it 30 % slower in the software implementations. This was achieved by comparing the running time of the protocol for 1 million encryptions.

Another block cipher where we can replace the diffusion layer by the proposed one is Hierocrypt [9]. Hierocrypt does not explicitly use big-size S-boxes, but it constructs 32 bit S-boxes by using nested SPN structure together with four 8-bit S-boxes and the MDS L matrix. For diffusion within those 32-bit S-boxes, a 16×16 binary matrix called MDS H is used, which is MDS for four 32-bit inputs. If we use our proposed diffusion layer with the same L(x), instead of the MDS H [9], we can achieve a 2 times faster implementation with the same level of security.

AES Mix-column layer has a simple implementation. As another comparison, we decided to replace the MDS H matrix in Hierocrypt with the MDS matrix of AES. But since MDS code of AES is over GF(28) and the inputs of MDS H are four 32-bit words, we modified the corresponding irreducible polynomial in AES and replaced it with x 32+x 7+x 5+x 3+x 2+x+1 [10] to work over GF(232), which would still remain MDS. We call this new construction, sch 1. As another construction, we replaced the MDS H in Hierocrypt with the MDS code we proposed above in our solution and we called it sch 2. We observed that sch 2 still brings on 5 % better performance compared to sch 1.

3 Other Desirable Structures for the Proposed Diffusion Layer

In Sect. 2, the general form of the proposed diffusion layer was introduced in Fig. 1. Then, by assuming a special case of α i ’s and β i ’s, an instance of this diffusion layer was given in Eq. (3). In this section, we obtain all sets of α i ’s and β i ’s such that the diffusion layer of Fig. 1 becomes perfect. We know some properties of α i ’s and β i ’s; for instance if all the words of the output are directly represented as a function of input words, a function of each x i (0≤is−1) must appear in each equation. Another necessary condition is obtained for two active words of the input. Assume there exist only two indices i,j such that x i , \(x_{j}\not= 0\). If we write each two output words y p , y q in a direct form as a function of x i and x j , we obtain

$$ \left\{ \begin{array}{l} y_{p}= L_{p_i}( x_i) \oplus L_{p_j}( x_j)\\ y_{q}= L_{q_i}( x_i) \oplus L_{q_j}( x_j)\\ \end{array} \right. $$

If

$$\frac{\mathcal{L}_{p_i}}{\mathcal{L}_{q_i}}=\frac{\mathcal {L}_{p_j}}{\mathcal{L}_{q_j}} \quad\mbox{or}\quad \left\lvert \begin{array}{c@{\quad }c} \mathcal{L}_{p_i} & \mathcal{L}_{p_j}\\ \mathcal{L}_{q_i} & \mathcal{L}_{q_j} \end{array} \right\lvert=0 $$

then, y p =0 is equivalent to y q =0. Thus, the minimum number of active words in the input and output is less than or equal to s and the branch number will not reach the maximal value s+1. This procedure must be repeated for 3, and more active words in the input. As an extension, we can use Lemma 3 of [10].

Lemma 8

([10])

Assume the diffusion layer has m inputs/outputs bits, and \(\mathcal {L}\) is the linear operator of L(x), and I is the linear operator of I(x). Moreover, ML D is an m×m matrix representation of the operator of the diffusion layer. If D is perfect, then all the submatrices of ML D are non-singular.

If we construct the ML D of Eq. (3), we have

$$\begin{aligned} \mathbf{ML}_D= \left(\begin{array}{c@{\quad}c@{\quad}c@{\quad}c} I & \mathcal{L} & I & I\oplus\mathcal {L}\\ I\oplus\mathcal{L} & I\oplus\mathcal{L}\oplus\mathcal {L}^2 & I & \mathcal{L}^2\\ \mathcal{L}^2 & I\oplus\mathcal{L}\oplus\mathcal{L}^3 & I\oplus\mathcal{L} & I\oplus\mathcal{L}^2\oplus \mathcal{L}^3\\ I\oplus\mathcal{L}^2\oplus\mathcal{L}^3 & \mathcal{L}\oplus \mathcal{L}^2\oplus\mathcal{L}^3\oplus\mathcal{L}^4 & \mathcal{L}\oplus\mathcal{L}^2 & \mathcal{L}^2\oplus\mathcal {L}^4 \end{array}\right) \end{aligned}$$

when calculating 69 sub-matrix determinants of ML D , we observe that these submatrices are non-singular only if L fulfills the condition of Theorem 7. However, by following this procedure, it is complicated to obtain all sets of α i ’s and β i ’s analytically. So, by systematizing the method based on Lemma 8, we performed a computer simulation to obtain all sets of α i ’s, and β i ’s in the diffusion layer in Fig. 1 that yield a perfect diffusion. We searched for all α i ’s and β i ’s that make the diffusion layer of Fig. 1 a perfect diffusion layer. This procedure was repeated for s=2,3,…,8. We found one set of (α i , β i ) for s=2, four sets for s=3, and four sets for s=4. The obtained diffusion layers along with the conditions on the underlying linear function L are reported in Table 2. We observed that for s=5, 6, 7, 8 the diffusion layer introduced in Fig. 1 cannot be perfect.

Table 2. Perfect regular recursive diffusion layers for s<8 with only one linear function L.

Note that some linear functions in Table 1 such as L(x (64))=(x (64)≪15)⊕(x (64)≫1) are not suitable for diffusion layers, since x (64)L 15(x (64)) must be invertible.

As we can see in Fig. 1, and its instances presented in Table 2, there exists some kind of regularity in the equations defining y i ’s, in the sense that the form of y i+1 is determined by the form of y i , and vice versa (F i ’s are all the same in Eq. (1)). However, we can present some non-regular recursive diffusion layers with a more general form (F i ’s are different) as in Fig. 3, where A i,j ,B i,j ∈{0,1}.

Fig. 3.
figure 3

Non-regular recursive diffusion layers.

If A i,j =α (ji)mods , and B i,j =β (ji)mods , then Fig. 3 is equivalent to Fig. 1. The main property of this new structure is that it still has one linear function L, and a simple structure for the inverse. For example, if s=4, then, the diffusion layer D is

$$ \left\{ \begin{array}{l} y_{0}= x_{0} \oplus A_{0,1} \cdot x_1 \oplus A_{0,2} \cdot x_2 \oplus A_{0,3} \cdot x_3 \oplus L(B_{0,1} \cdot x_1 \oplus B_{0,2} \cdot x_2 \oplus B_{0,3} \cdot x_3)\\ y_{1}= x_{1} \oplus A_{1,0} \cdot y_0 \oplus A_{1,2} \cdot x_2 \oplus A_{1,3} \cdot x_3 \oplus L(B_{1,0} \cdot y_0 \oplus B_{1,2} \cdot x_2 \oplus B_{1,3} \cdot x_3)\\ y_{2}= x_{2} \oplus A_{2,0} \cdot y_0 \oplus A_{2,1} \cdot y_1 \oplus A_{2,3} \cdot x_3 \oplus L(B_{2,0} \cdot y_0 \oplus B_{2,1} \cdot y_1 \oplus B_{2,3} \cdot x_3)\\ y_{3}= x_{3} \oplus A_{3,0} \cdot y_0 \oplus A_{3,1} \cdot y_1 \oplus A_{3,2} \cdot y_2 \oplus L(B_{3,0} \cdot y_0 \oplus B_{3,1} \cdot y_1 \oplus B_{3,2} \cdot y_2) \end{array} \right. $$

We searched the entire space for s=3 and s=4 (the order of search is 22s(s−1)). For s=3, we found 196 structures with branch number 4, and for s=4, 1634 structures with branch number 5. The conditions on linear functions that caused maximal branch number, are different for each structure. Among the 196 structures for s=3, the structure with the minimum number of operations (only 7 XORs, and one L evaluation) is the following:

$$ D: \left\{ \begin{array}{l} y_{0}= x_{0} \oplus x_1 \oplus x_2 \\ y_{1}= x_{1} \oplus x_2 \oplus L( y_0 \oplus x_2 )\\ y_{2}= x_{2} \oplus y_0 \oplus y_1 \end{array} \right. $$

where L(x) and xL(x) must be invertible.

This relation is useful to enlarge the first linear function of the hash function JH for 3 inputs [14]. For s=4, we did not find any D with the number of L evaluations less than four. However, the one with the minimum number of XORs is given as below:

$$ D: \left\{ \begin{array}{l} y_{0}= x_{0} \oplus x_1 \oplus x_2 \oplus L(x_3)\\ y_{1}= x_{1} \oplus x_3 \oplus y_0 \oplus L( x_2 \oplus y_0 )\\ y_{2}= x_{2} \oplus x_3 \oplus y_0 \oplus L(x_3 \oplus y_1)\\ y_{3}= x_{3} \oplus y_1 \oplus y_2 \oplus L( y_0 ) \end{array} \right. $$

Searching the whole space for s=5,6,… is too time consuming (note that for s=5 the order of search has complexity 240), and we could not search all the space for s≥5.

4 Increasing the Number of Linear Functions

In Sect. 3, we observed that for s>4 we cannot design a regular recursive diffusion layer in the form of Fig. 1 with only one linear function L. In this section, we increase the number of linear functions to overcome the regular structure of the diffusion layer of Eq. (3). A new structure is represented in Fig. 4, where α k ,β k ,γ k ∈{0,1}, k∈{0,1,…,s−1}, α 0=1,β 0=0 and γ 0=0.

Fig. 4.
figure 4

Regular recursive diffusion layers with two linear functions L.

If L 1 and L 2 are two distinct linear functions, Fig. 4 is too complicated to easily obtain conditions on L 1 and L 2 that make it a perfect diffusion layer (the order of search for s input/output is 23(s−1)). To obtain simplified conditions for a maximal branch number, let L 1 and L 2 have a simple relation like \(L_{2}(x)=L_{1}^{2}(x)\) or \(L_{2}(x)=L_{1}^{-1}(x)\). For the linear functions in Eq. (8), L 2(x) is more complex in comparison to L(x). However, there exist some linear functions in the form of Eq. (9) such that L −1(x) is simpler than L 2(x). As an example, for L(x (32))=(x (32)x (32)≫31)⋘1, we have L −1(x (32))=((x (32)⋙1)⊕(x (32)⋙1)≫31), but L 2(x (32))=(x (32)⊕(x (32)≫31)⋘1)⊕((x (32)⋘1)≫31)⋘1.

In Table 3, we introduce some recursive diffusion layers with (L 1=L and L 2=L −1) or (L 1=L and L 2=L 2) that have maximal branch numbers. These diffusion layers are obtained similar to that of Table 2. In this table, for each case, only y 0 is presented. Other y i ’s can be easily obtained from Fig. 4, since the F i ’s are all the same.

Table 3. Some perfect regular diffusion layers for s=5,6,7,8 with two linear functions.

If the 14 linear functions:

$$\begin{array}{l@{\quad }l@{\quad }l} L(x) & I\oplus L(x) & I\oplus L^3(x)\\ I\oplus L^7(x) & I\oplus L^{15}(x) & I\oplus L^{31}(x)\\ I\oplus L^{63}(x) & I\oplus L^{127}(x) & I\oplus L^{255}(x)\\ I\oplus L^{511}(x) & I\oplus L^{1023}(x) & I \oplus L^{2047}\\ I\oplus L^{4095}(x) & I\oplus L^{8191}(x) \end{array} $$

are invertible (all irreducible polynomials up to degree 13), then all the diffusion layers introduced in Table 3 are perfect. One example for a 32-bit linear function satisfying these conditions is

$$L(x_{(32)})=\bigl(x_{(32)} \oplus(x_{(32)} \gg31)\bigr) \lll29 $$

5 Conclusion

In this paper, we proposed a new family of efficient diffusion layers (recursive diffusion layers) which are constructed using several rounds of Feistel-like structures whose round functions are linear. The proposed diffusion layers are very efficient and have simple inverses, thus they can be deployed to improve the security or performance of some of the current block ciphers and hash functions and in the design of the future lightweight block ciphers and hash functions, even providing provable security against differential and linear attacks. For a fixed structure, we determined the required conditions for its underlying linear function to be perfectly secure with respect to linear and differential attacks. Then, for the number of words in input (output) less than eight, we extended our approach, and found all the instances of the perfect recursive diffusion layers with the general form described in Fig. 1. Also, we proposed some other diffusion layers with non-regular forms. Finally, diffusion layers with two linear functions were proposed. By using two linear functions, we designed perfect recursive diffusion layers for higher number of words.