Introduction

Data envelopment analysis (DEA) is a nonparametric linear programming-based technique first developed by Charnes et al. (1978) for evaluating the performance of homogeneous decision-making units (DMUs) having multiple inputs and multiple outputs. Since then DEA has been implemented in various applications such as service quality evaluation (Najafi et al. 2015), evaluation of academic performance (Monfared and Safi 2013) and facility location problem (Razi 2018). In DEA, efficiency score of each unit is generally defined as the ratio of the weighted sum of outputs to the weighted sum of inputs. The corresponding input and output weights for each unit are then calculated by solving individual linear programming problems that maximize this ratio provided that the corresponding ratio for all the units does not exceed one. Units with efficiency score of one are submitted as efficient and lay on the efficient frontier of the technology, whereas a score less than one signals inefficiency. Zamani and Borzouei (2016) investigated the stability region for preserving this classification in the presence of variable returns to scale. On the other hand, one of the main functions of DEA is benchmarking. It means that for each inefficient unit, DEA determines a (virtual) target unit lying on the frontier that monitors the best levels of inputs and outputs for that unit to perform efficiently. The first research on this subject was developed by Frei and Harker (1999), who have investigated the issue of benchmarking based on projecting inefficient units onto the strongly efficient frontier of DEA, considering Euclidean distance. After that, some researchers generalized the idea of using Euclidean distance in order to define efficiency measure and also to project units onto the strongly efficient frontier, e.g., see Baek and Lee (2009), Amirteimoori and Kordrostami (2010), and Aparicio and Pastor (2014a, b). The issue of least distance to the frontier has been used in other applications of DEA, such as ranking units (e.g., see Ziari 2016; Aghayi and Tavana 2018). The concept of benchmarking based on least distance has also been developed by Pastor and Aparicio (2010), Ando et al. (2012, 2017), Aparicio and Pastor (2013, 2014a, b) and Aparicio et al. (2014). On the other hand, some authors like Cherchye and Van Puyenbroeck (2001) and Silva Portela et al. (2003) have developed some benchmarking approaches based in the issue of similarity and closeness. Their motivation was that a benchmark which is closer to the inefficient unit is easier to be reached. Other measures of efficiency, like the modified Russell measure and the slack-based measure, have been also utilized for benchmarking, e.g., see Gonzalez and Alvarez (2001) and Aparicio et al. (2007). However, benchmarking is considered as an important field of research in DEA and can be viewed from different perspectives such as neural network as discussed in Shokrollahpour et al. (2016), artificial units in Didehkhani et al. (2018) or Fuzzy De-Novo programming by Sarah and Khalili-Damghani (2018). For a complete review of other benchmarking models, the reader is referred to Aparicio (2016), Aparicio et al. (2017a, b).

An important feature of DEA in evaluating units is that it behaves naturally optimistic, in the sense that it primarily treats units as individual and independent entities, as the corresponding optimal weights for each unit are calculated by autonomous programs. Therefore, it would be likely to happen that the input and output weights differ considerably across all units, which is incompatible and irrelevant. Especially, flexibility in choosing each unit’s optimal weights will surely reduce the discrimination power of DEA in distinguishing efficient and inefficient units. This query usually causes critical arguments among researchers. In this regard, there are a variety of means to overcome the aforementioned difficulty. For an early review of methods on improving the power of discrimination in DEA see Angulo-Meza and Lins (2002). One basic approach suggested by Li and Reeves (1999) is the multiple criteria DEA (MCDEA) technique. They formulated DEA in the framework of multi-criteria decision making (MCDM) and tried to evaluate units applying MCDM techniques. Recently, Chaves et al. (2016) studied the main properties of the model proposed by Li and Reeves (1999). On the other hand, Bal and Orkcu (2007) formulated a goal programming problem for weight dispersion in DEA. Their approach was later improved by Bal et al. (2010). Furthermore, Ghasemi et al. (2014) suggested an approach for improving the discrimination power in MCDEA. This study was recently supplemented by Rubem et al. (2017) who introduced a weighted goal programming formulation to solve the MCDEA problem.

Another basic approach to improve the discrimination power of DEA is to determine a common set of weights for all units under assessment. This approach was first proposed by Roll et al. (1991). Actually, a common set of weights (CSWs) are basically considered as coefficients of a supporting hyperplane of the DEA technology at some efficient units. However, the question of how to determine such a hyperplane has been seen in different perspectives. One approach is to minimize the difference between the DEA efficiency scores and those obtained from the associated CSWs. This method was investigated by Despotis (2002) and Kao and Hung (2005). Other formulations have been also proposed, such as maximizing the sum of the efficiency ratios of all the units by Ganley and Cubbin (1992), maximizing the minimum of the efficiency ratios by Troutt (1997), or introducing ideal and anti-ideal virtual units by Khalili-Damghani and Fadaei (2018).

Additionally, Ruiz and Sirvant (2016) formulated a model of CSW in the framework of benchmarking, i.e., they formulated a program to globally minimize a weighted \(L_1\)-distance of all the units to their corresponding benchmarks which are located on a common hyperplane of the DEA technology. In this regard, they suggested that only a facet of the DEA technology should be considered as the best practice frontier. Then, the coefficients of this hyperplane are implemented to calculate efficiency scores. Therefore, it is reasonable to establish benchmarks on this frontier. However, it should be noted that applying this method for target setting may involve deterioration of some of the observed input and output levels. In other words, in the common benchmarking approach the dominance criteria do not prevail necessarily. Moreover, this approach provides benchmark activities for all units under evaluation, simultaneously. Even for efficient units which are not located on the common best practice frontier, a different benchmark unit may be assigned, located on the underlying hyperplane. The inquiry that is highlighted here is that as all the benchmarks lay on a common supporting hyperplane of the technology, it is likely happen that the determined best practice efficient target may be very far from the corresponding unit. Therefore, it would be quite impossible for that unit to achieve its (final) target in one single step, because large-scale input and output adjustments are quite problematic and very demanding for an inefficient unit. This issue is also of great importance even if the final target is obtained via conventional DEA model for individual units separately. In this regard, some researchers investigated the problem of stepwise target setting. This issue includes developing a set of intermediate targets which constitute a path toward the strongly efficient frontier on which the final target is located. The methods of stepwise target setting can be categorized into two types. The first category involves the approaches in which all intermediate targets and final target are chosen from among the set of observed (or existing) units. In contrast, all the approaches that allow all virtual units to be selected as intermediate and final target belong to the second category. In this regard, Seiford and Zhu (2003) were the first who developed a model for type one stepwise benchmarking. Their model has been formulated based on context dependent DEA and layering units into successive layers. After that, most of the existing models in this category adapted the idea of this idea, e.g., Lim et al. (2011) established a path of intermediate benchmarks by clustering units and layering them. Lozano and Villa (2005a, b) investigated the concept of efficiency improvement in DEA, which leads to a stepwise target setting model. Lozano and Villa (2005a) proposed a model which is based on bounded adjustments of inputs and outputs, such that only a limited portion of the corresponding input and output levels is allowed for reduction and expansion at each step, respectively. Also, their model guarantees that each of intermediate targets dominates the previous one. Finally, the algorithm terminates when an efficient target is reached. Also, Suzuki and Nijkamp (2011) developed a stepwise projection DEA model for public transportation in Japan. Khodakarami et al. (2014) formulated a two-stage DEA model which has been applied on industrial Parks. Additionally, Fang (2015) developed a centralized resource allocation based on gradual efficiency improvement. For a complete review on the stepwise target setting models see Lozano and Calzada-Infante (2017). Moreover, Nasrabadi et al. (2018) investigated the concept of stepwise benchmarking in the presence of interval scale data in DEA.

The above discussion clarifies that the enquiry to investigate the problem of sequential benchmarking in case that all final benchmarks are located on a common hyperplane, is of great importance in DEA. In this study, we aim to investigate this subject in the framework of the CSW proposed by Ruiz and Sirvant (2016). Also, we incorporate the idea of bounded adjustments of Lozano and Villa (2005a, b) in our proposed method. Recalling that in the approach of Ruiz and Sirvant (2016) even efficient units that do not lay on the common best practice frontier are also assigned with a benchmark, a qualified model for determining a sequence of benchmarks should be applicable for both efficient and inefficient units. Therefore, we enquire about these two cases separately and establish different formulations for each one, although both models are basically similar.

This paper unfolds as follows. In “A common benchmarking model in DEA” section, we present some preliminary considerations of DEA and review the common benchmarking model developed by Ruiz and Sirvant (2016). Our proposed approach for sequential benchmarking is illustrated in “Developing a benchmark path toward the common best practice frontier” section, followed by a simple numerical example in “Numerical example” section. Then, “Application” section includes an empirical application and finally “Conclusion” section provides discussion and concluding remarks.

A common benchmarking model in DEA

In this section, we briefly present the common benchmarking model developed by Ruiz and Sirvant (2016). Consider a set of n decision-making units (DMUs), each consuming m inputs to produce s outputs. For \(j=1,\ldots ,n\), we denote unit j by the activity vector \((X_j,Y_j)\), where \(X_j\) and \(Y_j\) are input and output vectors of unit j, respectively. It is assumed that \(X_j=(x_{1j},\ldots ,x_{mj})^t \in {\mathcal {R}}^m_{\ge 0}\), \(X_j \ne 0\) and \(Y_j=(y_{1j},\ldots ,y_{sj})^t \in {\mathcal {R}}^s_{\ge 0}\), \(Y_j \ne 0\).

The production possibility set consisting all feasible input–output vectors (XY) is generally defined as \(T=\{(X,Y) | X\; {\text{ can }}\; {\text{ produce }}\;Y \}\). Each member of T is called a (virtual) activity. We call each unit j an observed activity or an observed unit, for \(j=1,\ldots ,n\).Footnote 1 An activity \((\bar{X},\bar{Y})\in T\) is said to be non-dominated or efficient in set T iff there exists no other \((X,Y)\in T\) such that \(X \le \bar{X}\) and \(Y \ge \bar{Y}\). The set of all observed extreme efficient units is denoted by E.

If we assume a variable returns to scale (VRS) technology, the set T in DEA is characterized as:

$$\begin{aligned} T_v=\left\{(X,Y)\ |\ X\ge \sum _{j=1}^n {\lambda _j X_j}, Y\le \sum _{j=1}^n {\lambda _j Y_j}, \sum _{j=1}^n{\lambda _j} =1, \lambda _j \ge 0, j=1,\ldots ,n \right\} \end{aligned}$$
(1)

All models here are formulated for a VRS technology. However, the set T for constant returns to scale (CRS) technology can be easily formulated by dropping the affine constraint \(\sum\nolimits _{j=1}^n{\lambda _j} =1\) from the above formulation.

The CSW model of Ruiz and Sirvant (2016)

Based on above notations, the common benchmarking model which provides the closest targets on a common efficient frontier of the technology for all units simultaneously, is formulated as (see Ruiz and Sirvant 2016):

$$\begin{array}{lll} \min &{} \quad \sum\limits_{j=1}^n{\Vert (X_j,Y_j)-(X_j^*,Y_j^*)\Vert ^{\mathbf {w}}_1} \\ {\text{s.t.}} &{} \quad X_j^* = \sum\limits_{k\in E} {\lambda _k^j X_k} &{} \quad j=1,\ldots ,n \\ &{} \quad Y_j^* = \sum\limits_{k\in E} {\lambda _k^j Y_k} &{} \quad j=1,\ldots ,n \\ &{} \quad \sum\limits_{k\in E} {\lambda _k^j} = 1 &{} \quad j=1,\ldots ,n \\ &{} \quad \lambda _k^j \ge 0 &{} \quad j=1,\ldots ,n, k\in E \\ &{} \quad X_j^* \ge 0, Y_j^* \ge 0 &{} \quad j=1,\ldots ,n \\ &{} \quad -U Y_k + V X_k + \gamma - d_k =0 &{} \quad k \in E \\ &{} \quad U \ge {\mathbf {w}}^{\mathbf {y}}, V \ge {\mathbf {w}}^{\mathbf {x}}, \gamma \; {\text{ free }} \\ &{} \quad d_k \le M b_k &{} \quad k \in E \\ &{} \quad \sum\limits_{j=1}^{n} {\lambda _k^j} \le M (1-b_k) &{} \quad k \in E \\ &{} \quad b_k \in \{ 0, 1 \}, d_k \ge 0 &{} \quad k \in E, \\ \end{array}$$
(2)

where \(\Vert (X_j,Y_j)-(X_j^*,Y_j^*)\Vert ^{\mathbf {w}}_1 = \sum _{i=1}^{m}{w^x_i |x_{ij}-x_{ij}^*|} + \sum _{r=1}^{s}{w^y_r |y_{rj}-y_{rj}^*|}\) is the weighted \(L_1\)-distance in \({\mathcal {R}}^{m+s}\) space, with the nonnegative weight vector \({\mathbf {w}}=({\mathbf {w}}^{\mathbf {x}},{\mathbf {w}}^{\mathbf {y}}) \in {\mathcal {R}}^{m+s}\). Also, M is assumed to be a sufficiently large positive quantity. Solving model (2), we have:

  1. 1.

    \(H^*=\{ (X,Y)\ |\ -U^* Y + V^* X + \gamma ^* =0 \}\) is the common best practice frontier of the technology which is used as the reference hyperplane in target setting.Footnote 2

  2. 2.

    The set \(RG=\{ k\ |\ \lambda _k^{j*} >0\;{\text{ for }}\;{\text{ some }}\;j, 1\le j\le n \}\) provides the reference group of units which lay on the reference hyperplane \(H^*\).

  3. 3.

    For unit j which is not in the reference group, the coordinates of the corresponding target on \(H^*\) is given by

    $$\begin{aligned} (X_j^*, Y_j^*)=\sum _{k \in E} {\lambda _k^{j*} (X_k, Y_k)}. \end{aligned}$$
    (3)

Note that model (2) can be easily transformed to a zero–one linear program by using transformations \(|x|=x^+ + x^-\) and \(x=x^+ - x^-\) for \(x \in {\mathcal {R}}\).

Difficulties in common benchmarking

Although model (2) determines benchmark activities for all units in the sense of closest targets, but the corresponding targets may be very far from the original units with actual input and output levels in practice. The main reason is that all targets are supposed to lay on a common supporting hyperplane of the underlying technology. Therefore, this makes it roughly impossible for some units to achieve the reference hyperplane in a single action, as it predictably requires significant adjustments of inputs and/or outputs. This issue is also of great importance even for efficient units which do not belong to RG. Here, we illustrate this point by a simple numerical example including 15 units with a single input and a single output. The original data set is provided by the first two columns of Table 1.

Table 1 Data set of the numerical example

In order to determine the CSWs, target activities and the common best practice frontier, we run model (2) with \(({\mathbf {w}}^{\mathbf {x}},{\mathbf {w}}^{\mathbf {y}})=({\mathbf {1}}_m,{\mathbf {1}}_s)\) and obtain \(H^*=\{ (x,y)\ |\ -2y+x+11=0 \}\) as the reference hyperplane. Also, the set \(RG=\{\mathbf{C },\mathbf{D },\mathbf{P }\}\) is determined as the index set of the reference group. Moreover, the corresponding targets are found out and given in the last two columns of Table 1. The corresponding technology in input–output space is represented in Fig. 1, and the common reference hyperplane is shown by bold line.

Fig. 1
figure 1

The production possibility set and the reference hyperplane for illustrative example

It is clear that all units in RG, i.e., units \(\mathbf{C }\), \(\mathbf{D }\) and \(\mathbf{P }\) coincide their corresponding targets. This also happens for the efficient non-extreme unit \(\mathbf{P }\) which is located on \(H^*\). Moreover, the common benchmark for efficient units \(\mathbf{A }\), \(\mathbf{B }\) and \(\mathbf{N }\) is unit \(\mathbf{C }\) and for unit \(\mathbf{D }\) is unit \(\mathbf{E }\). Additionally, the corresponding benchmark for the other inefficient units is a virtual (or observed) activity located on the intersection of \(H^*\) and \(T_v\).

Now, consider unit \(\mathbf{A }\). We observe that although it is efficient, but it has a different benchmark, i.e., unit \(\mathbf{C }\), which is approximately far from it. Therefore, it might be disappointing for unit \(\mathbf{A }\) to just consider such a far unit as its benchmark, since it is roughly impossible to reach it in a single move. Moreover, there exists a similar story for, e.g., the inefficient unit \(\mathbf{I }\).

This setting motivates us to address the above mentioned problem, i.e., to help each unit to achieve the common best practice frontier gradually. We follow a procedure to set up a path of intermediate benchmarks for each unit until we reach \(H^*\). The procedure is developed in next section.

Developing a benchmark path toward the common best practice frontier

In order to develop a procedure for establishing a benchmark path, efficient and inefficient units are investigated separately. Recalling that \(H^*\) is the reference hyperplane characterized by model (2) and assuming that unit “o” is under assessment, with \(o \notin RG\), we aim to establish a sequence of (improving) intermediate benchmarks for this unit that originates from it and converges to \(H^*\) in sequential steps. Let \((Target-t)\) be the model which calculates the tth (intermediate) target from the preceding one. Then, we present the general scheme of our benchmarking algorithm as the following 3-step procedure:


Sequential Target Setting Algorithm for unito

  • Step 1 Set \(t=0\), \((X_o^t,Y_o^t)=(X_o,Y_o)\).

  • Step 2 Set \(t=t+1\) and solve \((Target-t)\) for unit “o” to obtain

    $$\begin{aligned} \left(X_o^{t},Y_o^{t}\right)=\left(X_o^{t-1} - S^{-t*},Y_o^{t-1} + S^{+t*}\right). \end{aligned}$$
  • Step 3 If \((X_o^{t},Y_o^{t}) \in H^*\), stop. Otherwise go to Step 2.

Now, the main question is that how to formulate \((Target-t)\) for each unit. The answer is that the structure of \((Target-t)\) depends on the efficiency status of unit “o” under consideration. We first try to formulate \((Target-t)\) for an efficient unit, and then, we go to the case of an inefficient one.

A sequence of targets for an efficient unit

Assuming that unit “o” is efficient, we develop a model which settles intermediate benchmarks through a path toward \(H^*\). We acknowledge that the following remarks should be embedded in the proposed model:

  1. 1.

    As unit “o” is efficient, all intermediate benchmarks are expected to be efficient, too. This means that the benchmark path must go through the efficient frontier until it reaches \(H^*\), which is an especial part of the efficient frontier. This implies that the efficiency status of all intermediate benchmarks is not allowed to be deteriorated along the path. Nevertheless, the benchmark path would inevitably involve deterioration in some of the observed input and/or output levels in return to improving the others.

  2. 2.

    The path is monotonically convergent to \(H^*\), i.e., the direction of the path is toward the common best practice frontier. This means that each intermediate benchmark is closer to the final target than the previous one. We verify this property in our approach by checking whether each individual input/output of unit “o” has improvement in its final target or not. If an individual input (output) has been improved (deteriorated) in the final target, we imply that this property should be satisfied for all intermediate benchmarks. Therefore, the sign of each input/output slack in intermediate targets is determined according to the sign of the corresponding slack in the final target.

  3. 3.

    As the path is expected to move toward \(H^*\) gradually, it is rational to approve bounded adjustments in input and output levels at each step. Hence, like Lozano and Villa (2005a, b) we set upper and lower bounds for input and output adjustments. These bounds are defined as a predetermined portion of their current levels. Note that the corresponding bounds are determined by the decision maker, based on some managerial limitations and they may differ for individual inputs and outputs at each step.

All in all, the proposed benchmark model \((Target-t)\) associated with unit “o”, for \(t=1,2,\ldots\) is formulated as:

$$\begin{array}{llll} \max &{} \quad \Vert (S^{-t},S^{+t})\Vert ^{{\mathbf {w}}}_1 &{}&{}\quad (a) \\ s.t. &{} \quad X_o^{t-1} - S^{-t} = \sum\limits _{k\in E} {\lambda _k X_k} &{}&{}\quad (b) \\ &{} \quad Y_o^{t-1} + S^{+t} = \sum\limits _{k\in E} {\lambda _k Y_k} &{}&{}\quad (c) \\ &{} \quad \sum\limits _{k\in E} {\lambda _k} = 1 &{}&{}\quad (d) \\ &{} \quad \lambda _k \ge 0 &{}\quad k\in E &{} \quad (e) \\ &{} \quad -U Y_k + V X_k + \gamma - d_k =0 &{} \quad k \in E &{} \quad (f) \\ &{} \quad U \ge {\mathbf {w}}^{\mathbf {y}}, V \ge {\mathbf {w}}^{\mathbf {x}}, \gamma \; {\text{ free }} &{}&{}\quad (g) \\ &{} \quad d_k \le M b_k &{} \quad k \in E &{} \quad (h) \\ &{} \quad \lambda _k \le M (1-b_k) &{} \quad k \in E &{} \quad (i) \\ &{} \quad b_k \in \{ 0, 1 \}, d_k \ge 0 &{} \quad k \in E &{} \quad (j) \\ &{} \quad (x_{io}^{t-1} - x_{io}^*) s_i^{-t} \ge 0 &{} \quad i=1,\ldots ,m &{} \quad (k) \\ &{} \quad (y_{ro}^* - y_{ro}^{t-1}) s_r^{+t} \ge 0 &{} \quad r=1,\ldots ,s &{} \quad (l) \\ &{} \quad |s_i^{-t}| \le \delta _x x_{io}^{t-1} &{} \quad i=1,\ldots ,m &{} \quad (m) \\ &{} \quad |s_r^{+t}| \le \delta _y y_{ro}^{t-1} &{} \quad r=1,\ldots ,s &{} \quad (n) \\ &{} \quad |s_i^{-t}| \le | x_{io}^{t-1} - x_{io}^* | &{} \quad i=1,\ldots ,m &{} \quad (o) \\ &{} \quad |s_r^{+t}| \le | y_{ro}^{t-1} - y_{ro}^* | &{} \quad r=1,\ldots ,s. &{} \quad (p) \\ \end{array}$$
(4)

Note that, similar to the CSW model of Ruiz and Sirvant (2016), we use weighted \(L_1\)-norm in the objective function of the above model. To illustrate the model, note that constraints (b)–(j) guarantee that the tth target \((X_o^{t-1}-S^{-t},Y_o^{t-1} - S^{+t})\) lies on the efficient frontier (see model (2)). The next two constraints (k) and (l) determine the sign of slacks \(s_i^{-t}\) and \(s_r^{+t}\), respectively, depending on that input i and output r has improvement in the final target or not. This will ensure us that the constructed path moves uniformly toward the final target \((X_o^*,Y_o^*)\). In addition, two constraints (m) and (n) impose bounds on the value of slacks as adjustment values, where \(\delta _x\) and \(\delta _y\) are quantities between zero and one and denote relative amount of each input and output that is allowed for adjustment, respectively. Note that these values may vary in different steps. Finally, the last two constraints (o) and (p) are added to guarantee the convergence of the obtained path in the sense of \(L_1\)-distance. This fact is verified in the following theorem.

Theorem 1

Assume that\((X_o^{t},Y_o^{t})=(X_o^{t-1}-S^{-t*},Y_o^{t-1} - S^{+t*})\)is thetth benchmark for unitodetermined by model\((Target-t)\), and \((X_o^*,Y_o^*)\)is the final target obtained from model (2). Then:

$$\begin{aligned} \Vert (X_o^{t},Y_o^{t})-(X_o^*,Y_o^*)\Vert ^{\mathbf {w}}_1\le \Vert (X_o^{t-1},Y_o^{t-1})-(X_o^*,Y_o^*)\Vert ^{\mathbf {w}}_1. \end{aligned}$$
(5)

Proof

We have

$$\begin{aligned}\left\Vert \left(X_o^{t},Y_o^{t}\right)-\left(X_o^*,Y_o^*\right)\right\Vert ^{\mathbf {w}}_1 & = \left\Vert \left(X_o^{t-1}- S^{-t},Y_o^{t-1} + S^{+t}\right)-\left(X_o^*,Y_o^*\right)\right\Vert ^{\mathbf {w}}_1 \\ &= \sum _{i=1}^m{w^x_i \left|x_{io}^{t-1} - x_{io}^* - s_i^{-t}\right|} + \sum _{r=1}^s{w^y_r \left|y_{ro}^* - y_{ro}^{t-1} - s_r^{+t}\right|} \end{aligned}$$

Now, recalling that the weight vectors \({\mathbf {w}}^{\mathbf {x}}\) and \({\mathbf {w}}^{\mathbf {y}}\) are both nonnegative, it is sufficient to prove that \(|x_{io}^{t-1} - x_{io}^* - s_i^{-t}| \le |x_{io}^{t-1} - x_{io}^*|\) for \(i=1,\ldots ,m\), and \(|y_{ro}^* - y_{ro}^{t-1} - s_r^{+t}| \le |y_{ro}^* - y_{ro}^{t-1}|\) for \(r=1,\ldots ,s\). Toward this end, we consider three cases as:


Case 1: \(x_{io}^{t-1} - x_{io}^* >0\). In this case, constraint (k) implies that \(s_i^{-t} \ge 0\). On the other hand, by constraint (o), we have \(s_i^{-t} \le x_{io}^{t-1} - x_{io}^*\). Therefore, we have:

$$\begin{aligned} |x_{io}^{t-1} - x_{io}^* - s_i^{-t}| = x_{io}^{t-1} - x_{io}^* - s_i^{-t} \le x_{io}^{t-1} - x_{io}^* = |x_{io}^{t-1} - x_{io}^*|. \end{aligned}$$

Case 2: \(x_{io}^{t-1} - x_{io}^* <0\). First, by constraint (k), we have \(s_i^{-t} \le 0\). Meanwhile, constraint (o) implies that \(s_i^{-t} \ge x_{io}^{t-1} - x_{io}^*\). Therefore, we have:

$$\begin{aligned} |x_{io}^{t-1} - x_{io}^* - s_i^{-t}| = s_i^{-t} - (x_{io}^{t-1} - x_{io}^*) \le - (x_{io}^{t-1} - x_{io}^*) = |x_{io}^{t-1} - x_{io}^*|. \end{aligned}$$

Case 3: \(x_{io}^{t-1} - x_{io}^* =0\). In this case, constraint (o) implies that \(s_i^{-t} = 0\). Therefore, the inequality clearly holds by equality.

The above three cases prove the first aforementioned statement. By a similar discussion on outputs, one can prove the second inequality, easily. \(\square\)

As we wish to establish a benchmark path converging to the common best practice frontier, we run model (4) until we reach the common frontier \(H^*\), i.e., we have \(-U^* Y_o^{t} + V^* X_o^{t} + \gamma ^* =0\), where \((U^*,V^*,\gamma ^*)\) is an optimal solution of model (2).

A sequence of targets for an inefficient unit

Now, we assume that unit “o” is inefficient. In order to develop a mathematical formulation to establish intermediate benchmarks for this unit, we embed the following considerations in our model:

  1. 1.

    As unit “o” is inefficient, we aim to find its sequential benchmarks such that each intermediate benchmark is better that the previous one. Although in common benchmarking model (2) the domination criteria does not necessarily prevail, it is possible to determine intermediate benchmarks at each step in a way that each target is better than the previous one, in the sense of its weighted \(L_1\)-distance to the efficient frontier. Hence, we consider the following weighed additive model of Cooper et al. (2011) evaluating (virtual) activity \((\bar{X},\bar{Y})\in T\), which minimizes the weighted \(L_1\)-distance of this unit to the efficient frontier, as follows:

    $${\text{ADD}}(\bar{X},\bar{Y}) =\begin{array}{lll}\min & \quad -U \bar{Y} + V \bar{X}+\gamma &\\ {\hbox{s.t.}} &\quad -U Y_j + V X_j+\gamma \ge 0, &\quad j\in E \\ &{} \quad U \ge {\mathbf {w}}^{\mathbf {y}},\quad V \ge {\mathbf {w}}^{\mathbf {x}}, \gamma \;{\text{free}}, \\ \end{array}$$
    (6)

    where \(({\mathbf {w}}^{\mathbf {x}},{\mathbf {w}}^{\mathbf {y}})\in {\mathcal {R}}^{m+s}\) is a nonnegative weight vector. Our aim is to determine \((X_o^{t},Y_o^{t})\) from the previous (intermediate) benchmark \((X_o^{t-1},Y_o^{t-1})\) with the property that

    $$\begin{aligned} {\text{ADD}}(X_o^{t},Y_o^{t}) \le {\text{ADD}}(X_o^{t-1},Y_o^{t-1}). \end{aligned}$$
    (7)

    In other words, we wish to move along an improving direction in our benchmarking procedure. The following theorem provides us to find a sufficient condition for such a direction.

Theorem 2

Assume that\((U_o^*,V_o^*,\gamma _o^*)\)is an optimal solution of the weighted additive model (6) evaluating unito”. Then for each vector\((S^-,S^+)\in {\mathcal {R}}^{m+s}\)such that\(U_o^* S^+ + V_o^* S^- \ge 0\), we have\({\text{ADD}}(X_o - S^-,Y_o + S^+) \le {\text{ADD}}(X_o,Y_o)\).

Proof

The proof is straightforward if we apply the additive model (6) for evaluating \((X_o - S^-,Y_o + S^+)\). By assumption, it can be verified that \((U_o^*,V_o^*,\gamma _o^*)\) is a feasible solution for model (6) evaluating \((X_o - S^-,Y_o + S^+)\), which implies that \({\text{ADD}}(X_o - S^-,Y_o + S^+) \le {\text{ADD}}(X_o,Y_o)\), and the proof is complete. \(\square\)

The above theorem verifies that if we add the constraint \(U_o^* S^+ + V_o^* S^- \ge 0\) to our benchmarking model, we can guarantee that the resulting path is an improving one w.r.t. the weighted additive model (6).

  1. 2.

    It should be kept in mind that a desirable benchmark path should go monotonically toward the best practice frontier. So, as before, the sign of input and output slacks is determined according to improvement or deterioration in the corresponding input and outputs of the final target. This will be observed in the proposed model similar to model (4).

  2. 3.

    The final consideration deals with bounded adjustments for inputs and outputs at each step. This issue can be observed in the proposed model similar to model (4).

Finally, model \((Target-t)\) associated with inefficient unit “o” is formulated as:

$$\begin{array}{llll} \max &{} \quad \Vert (S^-,S^+)\Vert ^{\mathbf {w}}_1 &{}&{}\quad (a) \\ s.t. &{} \quad X_o^{t-1} - S^{-t} = \sum\limits_{k\in E} {\lambda _k X_k} &{}&{}\quad (b) \\ &{} \quad Y_o^{t-1} + S^{+t} = \sum\limits_{k\in E} {\lambda _k Y_k} &{}&{}\quad (c) \\ &{} \quad \sum\limits_{k\in E} {\lambda _k} = 1 &{}&{}\quad (d) \\ &{} \quad \lambda _k \ge 0 &{}\quad k\in E &{} \quad (e) \\ &{} \quad U_o^* S^+ + V_o^* S^- \ge 0 &{}&{}\quad (f) \\ &{} \quad (x_{io}^{t-1} - x_{io}^*) s_i^{-t} \ge 0 &{} \quad i=1,\ldots ,m &{} \quad (g) \\ &{} \quad (y_{ro}^* - y_{ro}^{t-1}) s_r^{+t} \ge 0 &{} \quad r=1,\ldots ,s &{} \quad (h) \\ &{} \quad |s_i^{-t}| \le \delta _x x_{io}^{t-1} &{} \quad i=1,\ldots ,m &{} \quad (i) \\ &{} \quad |s_r^{+t}| \le \delta _y y_{ro}^{t-1} &{} \quad r=1,\ldots ,s &{} \quad (j) \\ &{} \quad |s_i^{-t}| \le |x_{io}^{t-1} - x_{io}^*| &{} \quad i=1,\ldots ,m &{} \quad (k) \\ &{} \quad |s_r^{+t}| \le |y_{ro}^{t-1} - y_{ro}^*| &{} \quad r=1,\ldots ,s. &{} \quad (l) \\ \end{array}$$
(8)

To illustrate the model, note that constraints (b)–(e) guarantee that the tth benchmark is feasible. Constraint (f) ensures that the obtained benchmark is better than the previous one, w.r.t. the weighted additive model (6). Moreover, constraints (g) and (h) determine the direction of the path that should go toward the final target \((X_o^*,Y_o^*)\). Finally, the remaining constraints (g)–(l) are interpreted similar to what we have in model (4).

Computational complexity

Both models (4) and (8) that have been developed for sequential target setting, are nonlinear due to existence of absolute value in their constraints, as well as their objective functions. Nevertheless, this issue can be resolved by introducing two nonnegative variables \(x^+\) and \(x^-\), such that \(x=x^+ - x^-\) and \(|x|=x^+ + x^-\), but it increases the computational complexity of the models, which is not of interest and should be avoided. In this regard, we propose another approach for linearizing these models which is more practical. Without loss of generality, we illustrate this technique for model (4).

Assume that in model (4) we partition the set of input indices \(I=\{1,\ldots ,m\}\) as \(I = I^{t-1}_< \cup I^{t-1}_> \cup I^{t-1}_0\), where:

$$\begin{aligned} I^{t-1}_<& = \{ i\ |\ x_{io}^{t-1} - x_{io}^* < 0 \}, \\ I^{t-1}_> &= \{ i\ |\ x_{io}^{t-1} - x_{io}^* > 0 \}, \\ I^{t-1}_0 &= \{ i\ |\ x_{io}^{t-1} - x_{io}^* = 0 \}. \end{aligned}$$
(9)

Then, by constraints (k), (l) and (o), it is concluded that:

$$\begin{aligned} s_i^{-t} &\le 0, \quad i\in I^{t-1}_<, \\ s_i^{-t} &\ge 0, \quad i\in I^{t-1}_>, \\ s_i^{-t} &= 0, \quad i\in I^{t-1}_0. \end{aligned}$$
(10)

A similar technique can be used for output indices, easily. Now, according to (10), and by imposing suitable constraints for both inputs and outputs, one can easily ignore the absolute value notation in the corresponding models. Therefore, both models (4) and (8) are transformed to equivalent linear programs, accordingly.

Numerical example

Recall data set presented in Table 1. We have already obtained the final targets for all units by solving model (2). Now, we aim to set up a benchmark path for each individual unit, except for the RG ones. Based on our theory, we analyze efficient and inefficient units, by running models (4) and (8), respectively. Note that in both models we assume \(({\mathbf {w}}^{\mathbf {x}},{\mathbf {w}}^{\mathbf {y}})=({\mathbf {1}}_m,{\mathbf {1}}_s)\), and \(\delta _x = \delta _y = 0.3\). The results are reported in Table 2.

Table 2 Sequential target setting in the numerical example

From the results presented in Table 2, we observe that the three units \(\mathbf C\), \(\mathbf D\) and \(\mathbf P\) which belong to RG coincide their corresponding targets. Furthermore, the longest path is for unit \(\mathbf A\), which reaches its benchmarks in seven steps. On the other hand, unit \(\mathbf H\) has the shortest path to its benchmark, as it needs only one step to reach the reference hyperplane. Finally, we notice that the final destination of the benchmark path for all units is exactly the unit obtained from model (2). The benchmark path for all efficient and inefficient units, except for those belonging to RG, is shown in Fig. 2.

Fig. 2
figure 2

Target path for efficient and inefficient units

Application

To illustrate the proposed approach, consider the data taken from Coelli et al. (2002), which consists of 28 international airlines during year 1990. This data set has already been used in other DEA papers in order to illustrate different concepts (see Ray 2004, 2008; Ruiz 2013; Aparicio et al. 2007). Especially in Ruiz and Sirvant (2016), this data set has been applied to illustrate the concept of CSW and also ranking of units. Here, we apply our stepwise target setting model on this data set. Note that for each airline, four inputs and two outputs are considered which are provided in Table 3. For further details on the data set, one may refer to Ruiz and Sirvant (2016).

Table 3 Data set of 28 international airlines

First, a conventional DEA model under the assumption of CRS shows that \(E=\{4,6,7,8,11,13,15,16,18\}\). Then, we apply model (2) in the framework of CRS and with weights defined as:

$$\begin{aligned} w_i^x &= \frac{1}{\bar{x}_i}, &{}\quad i=1,\ldots ,m, \\ w_r^y &= \frac{1}{\bar{y}_r}, &{}\quad r=1,\ldots ,s, \end{aligned}$$
(11)

where \(\bar{x}_i\), \(i=1,\ldots ,m\) and \(\bar{y}_r\), \(r=1,\ldots ,s\) are the averages of the corresponding inputs and outputs of all units. Note that this specification of the \(L_1\)-distance has already been used in the DEA literature (see Thrall 1996). Moreover, by using the weighted \(L_1\)-distance with Eqs. (11), model (2) becomes unit invariant.

By running the model (2), we obtain \(RG=\{4,8,11,18\}\). Then, in order to find the target path for each unit, which does not belong to RG, we run models (4) and (8), for efficient and inefficient units, respectively. Note that in both models we set \(\delta _x=\delta _y=0.3\). The obtained sequential targets for efficient and inefficient units are reported in Tables 4 and 5, respectively.

Table 4 Sequential benchmarking for efficient units: actual inputs/outputs and target path
Table 5 Sequential benchmarking for inefficient units: actual inputs/outputs and target path

We first observe that Table 4 does not provide target path airlines in RG, because for these units targets and actual inputs/outputs coincide. However, for the efficient units not in RG, a path of sequential targets toward the reference hyperplane is reported. We observe that all efficient units reach their corresponding final target in two or three steps. For QANTAS and SAUDIA, the targets are less demanding, but the other efficient airlines, i.e., LUFTHANSA, SWISSAIR and PORTUGAL, need considerable adjustments to reach their corresponding target. The maximum adjustment for these three is due to their first input which needs approximately 50% decrease from their actual levels. Also, note that there would be no change in the third input of LUFTHANSA, the second input and second output of SWISSAIR and the second output of PORTUGAL. On the other hand, as the results show the efficiency status of the efficient units do not deteriorate along the path, i.e., for efficient airlines, all intermediate (and final) targets have an optimal value of zero in model (6).

On the other hand, turning to inefficient airlines in Table 5, we observe different behaviors. Among these airlines, CATHAY has the longest path including 5 benchmarks toward its final target. This is due to the level of its third input which needs a decrement of approximately 84% from its actual level (3171 to 549.8). Meanwhile, the level of its first input and second output remains unchanged along the path. Then, the second longest path among the inefficient airlines is of length 3 and belong to EASTERN and USAIR, both consisting of an adjustment of nearly twice in the second output. Although, three inefficient airlines reach their corresponding final target in two steps, the majority of ten inefficient airlines achieve the reference hyperplane in just one step. Moreover, the last column in Table 5 presents the optimal value of the weighted additive model (6) for each (virtual) unit. Observing these values, which constitute a descending sequence for each inefficient unit with a final value of zero, the improving characteristic of the obtained sequential benchmarks is confirmed.

Conclusion

One of the main features of DEA is that it can be used for benchmarking, which is an important issue in management and economics. In practice, in a production technology the decision maker (DM) usually aims to first evaluate the efficiency status of decision-making units and classify them into efficient and inefficient categories, and then to determine a benchmark feasible and efficient activity for each inefficient unit. Benchmarking can assist inefficient units to improve their performance in comparison to best practices of others. On the other hand, applying an aggregated DEA-based model which finds a common set of weights to evaluating the efficiency score of all units, simultaneously, it is possible to determine a common supporting hyperplane of the technology as the best practice frontier. Thus, one could evaluate the units by means of the coefficients of this frontier and also to establish targets for all of them, on this common best practice frontier. However, in practice these targets may be difficult to reach in a single step, and therefore, it is required to propose a gradual improvement strategy to achieve the best practices. This research developed an approach of establishing a gradual improving path of targets for each unit which is not located on the best practice frontier. To apply the proposed method, the common set of weights (CSW) model of Ruiz and Sirvant (2016) is solved in order to find the final target for all units, simultaneously. Then for each unit, a path of targets is found which originates from that unit and proceeds gradually to the final efficient target which has been already determined. The obtained path is an improving one, in the sense that if the underlying unit is an efficient one, then all of the intermediate targets are efficient as well, and for inefficient units each of the intermediate targets is closer to the final target than the previous one, and also has a better performance than the previous one, in the framework of the weighted additive model of Cooper et al. (2011). Also, in the proposed model only a limited amount of adjustment is allowed. The portion of the current levels of inputs and outputs that is allowed for adjustment is determined by the DM based on his/her managerial points of view and practical limitations. Then, this feature guarantees that the path is roughly acceptable by the DM, as it is more practical and understandable.

This approach can be extended along different lines. One can investigate this issue for a free disposal hull (FDH) technology. Furthermore, the question of establishing a target path which in convergent to the final target in a predetermined number of steps might be interesting. Another possibility is to develop a procedure of stepwise target setting in the framework of other common weight methodologies, such as models proposed in Despotis (2002) or Kao and Hung (2005).