1 Introduction

Amid the COVID-19 pandemic globally, healthcare construction projects have received special attention from the governments in various countries. In such circumstances, the speed of construction was recognized as one of the most critical factors because of the limited capacity of hospitals to admit patients [1]. However, various factors are essential in the performance of healthcare construction projects, and it is crucial to determine the performance of the healthcare construction projects to identify the current weaknesses and solve them for future projects. It is worth mentioning that performance measurement is not an individual activity; it is a teamwork activity to consider the stakeholders’ opinions, especially about the importance of the performance indicators. It can be challenging to determine the performance of healthcare construction projects based on a large number of experts and requires a powerful decision support system to handle such a problem. This question may arise how the performance of the healthcare construction projects can be calculated based on the opinions of a large group of stakeholders? How can this activity be simplified to decrease the computational time and cost? These research questions will be addressed in the current study carefully.

Chan et al. [2] proposed a conceptual framework to illustrate the critical performance factors of construction projects. They have considered five major groups, including project procedures, external environment, project-related factors, human-related factors, and project management actions, as important performance factors in construction projects. Jha and Iyer [3] identified 55 criteria associated with the quality performance of the construction projects. With the aid of statistical analysis, they have divided the specified criteria into two major groups, namely performance and failure factors. As a result, they found that top management support and project managers’ competence are two significant factors that can enhance the construction project’s quality performance. Rojas and Kell [4] incorporated the metrics such as change orders, cost growth, and unit cost to compare the cost performance of the construction projects. The main limitation of their study was focusing on the building projects of the Pacific Northwest school between 1987 and 2005. Bilbo et al. [5] employed two project delivery methods, namely Integrated Project Delivery (IPD) and Construction Manager at Risk (CMR), for healthcare facilities. They also addressed the advantages and challenges of implementing IPD and CMR methods. It was the first study that examined the performance of the IPD and CMR methods in healthcare projects and compared them together. Doulabi and Asnaashari [6] tried to identify the critical factors in the performance of the healthcare facility projects in Iran through interviews with experts with significant experience in this field of study. Finally, they identified eight groups of factors, including scope, time, cost, quality, human resource, risk, environmental, and external matters. It should be noted that 20 experts have identified these factors with at least 15 years of working experience. Hanna [7] evaluated the impact of IPD on the performance of building construction projects by utilizing a broad range of performance attributes. This study used quantitative analyses, which distinguished it from other existing research studies in the same field of study. Iskandar et al. [8] proposed a mathematical model to extract the weights of the performance attributes to determine the performance of the healthcare construction projects objectively. However, no study considers the performance measurement of the construction projects as a large-scale group decision-making (LSGDM) problem. There are several stakeholders in each project, and their opinions should be considered in the performance measurement process.

There are a few studies that utilized LSGDM in the healthcare management context. Liu et al. [9] proposed a new risk priority model utilizing prospect theory and cluster analysis, where there was a large number of experts. They used the similarity measure-based clustering method and considered conflict degree and majority principle, which could enhance the consistency of the experts’ opinions. Also, an entropy-based method was developed to measure the weight of risk factors that could determine the weights objectively. Liu et al. [10] proposed a large group dependence assessment method utilizing interval 2-tuple linguistic variables as well as cluster analysis to incorporate the dependence in human reliability analysis. They also improved the Muirhead mean operator so as to measure the dependence levels among operator actions. Li and Wei [11] presented an LSGDM framework for decision-making in healthcare management. In their study, the input data were hesitant fuzzy linguistic terms, and a novel clustering approach was proposed to categorize the decision-makers into clusters. Zhang and Meng [12] proposed an LSGDM method for selecting mobile health applications in hospitals. They considered both non-cooperative and cooperative behaviours of the experts in their study. To the best of our knowledge, no study used LSGDM for performance measurement in the healthcare management context. Moreover, this is the first study that extends the Ordinal Priority Approach (OPA) under LSGDM in healthcare management systems.

Based on the aforementioned studies, the contribution of the current study can be summarized as follows:

  • Pioneering a novel framework for determining the weight of attributes in LSGDM problems using the OPA. Moreover, proposing an equally-weighted OPA model and group-weighted OPA model, which can be used to decrease the computational complexity in special cases for solving group decision-making problems.

  • Proposing Weighted Kendall’s W to measure the confidence level in LSGDM problems, also proposing a formula to estimate computational complexity in the OPA model.

  • Proposing a framework to determine the performance of the healthcare construction projects using big data for the first time.

The rest of the study is organized as follows: after the introduction section and addressing the relevant studies in the LSGDM context, the preliminaries and methodology section are provided. Then, the proposed decision support system is explained. Afterwards, a comparative analysis is presented to show the feasibility of the proposed framework. Next, a case study that includes five healthcare construction projects is discussed, and the performance of the projects is determined using the proposed approach. Later, a comprehensive sensitivity analysis is performed on the case study to check the stability of the solutions. Finally, the conclusion and discussion are provided.

2 Large-scale group decision-making

This section aims to review the LSGDM context in the recent years. Liu et al. [13] proposed a decision-making framework for complex LSGDM problems in the intuitionistic fuzzy environment using partial least squares (PLS) path modeling. They addressed the application of the framework to the site selection of hydropower stations in China. Palomares et al. [14] presented a consensus model which was suitable for a large number of experts in decision-making problems using visual analysis tools. They believed that their model could identify non-cooperative behaviours in LSGDM problems using a fuzzy clustering-based approach. Liu et al. [15] proposed a decision-making framework to select the suitable contractor for mega construction projects using two-stage PLS path modeling. They also presented a case study in China and compared the typical PLS and two-stage PLS path modeling, which showed better reliability for two-stage PLS path modeling. Quesada et al. [16] presented a methodology based on uninorm aggregation operators to handle non-cooperative behaviours in LSGDM problems. They utilized computing with words and fuzzy theory approaches in the proposed weighting scheme to facilitate the usage of uninorm operators. Liu et al. [17] proposed a novel method for solving LSGDM problems using both subjective and objective viewpoints to calculate the groups’ weights. They believed that using percentage distribution can make the decision-making process more convenient and promote the ability of visualization. Xu et al. [18] developed a model for solving LSGDM problems under incomplete preference data. Their method was established based on subjective and objective trust degrees to convert the incomplete decision matrix to the complete decision matrix. Xiang [19] proposed a web-based approach to provide an automated platform for LSGDM under emergency conditions due to energy shortage while the time was also limited. Indeed, the experts could register on the platform and provide their opinions about the shortage of energy and the alternatives. These opinions could be used for the LSGDM to choose the best alternative in the emergency condition for the energy shortage. Zhang et al. [20] proposed a new computational model using linguistic information to solve the LSGDM problems. They addressed the application of the proposed model for the talent selection problem in universities. They extended various distance measurement and ranking methods to provide this approach. Rodríguez et al. [21] proposed a novel adaptive consensus reaching processes framework to solve LSGDM problems. This framework was established based on three important steps, including (i) clustering to determine the weight of the experts’ sub-groups, (ii) aggregating the opinions using the hesitant fuzzy approach to minimize information loss, (iii) providing a feedback mechanism, and evaluating the level of agreement. Song and Li [22] presented a model for solving LSGDM problems to attain the maximum level of information collected from the experts. They also employed extended Technique for Order Performance by Similarity to Ideal Solution (TOPSIS) for ranking the alternatives. Zhang [23] proposed a probabilistic method for solving LSGDM problems with probabilistic linguistic term sets. Moreover, a novel probabilistic averaging operator and distance measure formula were proposed to aggregate the experts’ opinions. Ding et al. [24] presented an intuitionistic fuzzy clustering method based on sparse representation to solve LSGDM problems. The main innovation of the model was investigating the intra-relations among the experts to find the leader of the group in the clustering process. Ma et al. [25] proposed a clustering method for solving LSGDM problems under hesitant fuzzy linguistic term sets. The proposed clustering method was advantageous in breaking large-scale problems into small problems. They also provided a new selection process to find the final solution to the problem. Wu et al. [26] stated that the experts’ opinions contained subjective uncertainty, to handle this uncertainty, they used interval type-2 fuzzy linguistic terms to solve LSGDM problems. They also proposed type-2 fuzzy k-means clustering to classify the experts’ opinions.

For more information, the applications of the LSGDM are summarized in Table 1 and were compared with the current study.

Table 1 A brief overview of selected studies involving LSGDM

Although there are several research items on the LSGDM context, they have not yet employed LSGDM for performance measurement. Moreover, the OPA is a new multiple attributes decision-making (MADM) method, and this is the first extension of it in LSGDM. Indeed, the current study aims to propose a performance measurement framework based on LSGDM and with the aid of the OPA for the first time. Also, Weighted Kendall’s W is proposed for the first time to calculate the confidence level in LSGDM considering the weight of the clusters.

3 Preliminaries

In this section, we aim to review and propose the essential tools and techniques, which should be employed in the proposed decision support system. First, the OPA of MADM is discussed, and some extensions are provided. Afterward, the ordinal K-means is presented, and the silhouette coefficient is explained. Finally, Weighted Kendall’s W is proposed to develop a confidence level index for the LSGDM problems.

3.1 Ordinal priority approach (OPA)

The OPA is a new MADM method proposed by Ataei, Mahmoudi, Feylizadeh, and Li [27]. This method requires ordinal data as input data to determine the weights of experts, attributes, and alternatives. The current study aims to utilize the OPA for solving LSGDM problems to assess the performance of healthcare construction projects. The sets, indexes, variables, and parameters are provided as follows, which are essential for a better understanding of the succeeding models.

Sets

I:

Set of clusters ∀i ∈ I

I′:

Set of experts ∀i' ∈ I'

J:

Set of attributes ∀j ∈ J

Indexes

i :

Index of the clusters (1, …, K)

i :

Index of the experts (1, …, p)

j :

Index of the attributes (1, …, n)

Variables

Z, Z′:

Objective function

\( {W_{i\prime j}}^{r_j} \) :

Weight (importance) of jth attribute based on expert’s opinion i at rank rj

Parameters

r i :

The rank of expert i

r i :

The rank of cluster i

r j :

The rank of attribute j

p i :

The number of experts with the same opinions in cluster i

The steps of the original OPA are explained as follows:

  • Step 1: Specifying the critical indicators which are necessary for performance measurement in the related healthcare construction project.

  • Step 2: Collecting data regarding the attributes from the experts who are the stakeholder in the healthcare construction project.

  • Step 3: Preparing Model (1) by utilizing the collected data in Steps 1 and 2. Then, this model can be solved using LINGO software.

$$ {\displaystyle \begin{array}{c}\mathit{\operatorname{Max}}\ Z\\ {}\begin{array}{c}S.t:\\ {}Z\le {r}_{i^{\prime }}\left({r}_j\left({W_{i^{\prime }j}}^{r_j}-{W_{i^{\prime }j}}^{r_j+1}\right)\right)\kern1.75em \forall {i}^{\prime } and\ {r}_j\ \\ {}\begin{array}{c}Z\le \kern0.5em {r}_{i^{\prime }}{r}_n{W_{i^{\prime }j}}^{r_n}\kern8.5em \forall {i}^{\prime } and\ {r}_j={r}_n\\ {}\sum \limits_{i^{\prime }=1}^p\sum \limits_{j=1}^n{W}_{i\prime j}=1\ \\ {}{W}_{i\prime j}\ge 0\kern12.5em \forall {i}^{\prime }\ and\ j\end{array}\end{array}\end{array}} $$
(1)
  • Step 4: By solving Model (1), the weights of the attributes and experts can be determined using Eqs. (2) and (3), respectively.

$$ {W}_j=\sum \limits_{i^{\prime }=1}^p{W}_{i\prime j}\kern3.5em \forall j $$
(2)
$$ {W}_{i\prime }=\sum \limits_{j=1}^n{W}_{i\prime j}\kern3.5em \forall {i}^{\prime } $$
(3)

Theorem 1

When the weights of the experts are considered equally \( \left(\raisebox{1ex}{$1$}\!\left/ \!\raisebox{-1ex}{$p$}\right.,\forall {i}^{\prime}\right) \), the optimum weights of Model (4) are equivalent to the optimum weights of Model (1) where ri = 1, ∀ i. Also, the optimum objective function of Model (4) is equal to the optimum objective function of Model (1) divided by p. Model (4) is called the equally-weighted OPA model.

$$ {\displaystyle \begin{array}{c}\mathit{\operatorname{Max}}\ {Z}^{\prime}\\ {}S.t:\\ {}\begin{array}{c}{Z}^{\prime}\le \left(\raisebox{1ex}{$1$}\!\left/ \!\raisebox{-1ex}{$p$}\right.\right)\left({r}_j\left({W_{i^{\prime }j}}^{r_j}-{W_{i^{\prime }j}}^{r_j+1}\right)\right)\kern1.75em \forall {i}^{\prime }\ and\ {r}_j\\ {}{Z}^{\prime}\le \kern0.5em \left(\raisebox{1ex}{$1$}\!\left/ \!\raisebox{-1ex}{$p$}\right.\right){r}_n{W_{i^{\prime }j}}^{r_n}\kern8.5em \forall {i}^{\prime }\ and\ {r}_j={r}_n\\ {}\begin{array}{c}\sum \limits_{i^{\prime }=1}^p\sum \limits_{j=1}^n{W}_{i\prime j}=1\ \\ {}{W}_{i\prime j}\ge 0\kern14em \forall {i}^{\prime }\ and\ j\end{array}\end{array}\end{array}} $$
(4)

Proof of theorem 1

Let us multiply the constraints of Model (4) by p, where p is the number of experts. Hence, Model (5) is obtained.

$$ {\displaystyle \begin{array}{c}\begin{array}{c}\mathit{\operatorname{Max}}\ {Z}^{\prime}\\ {}S.t:\\ {}{pZ}^{\prime}\le \left(\raisebox{1ex}{$p$}\!\left/ \!\raisebox{-1ex}{$p$}\right.\right)\left({r}_j\left({W_{i^{\prime }j}}^{r_j}-{W_{i^{\prime }j}}^{r_j+1}\right)\right)\kern1.75em \forall {i}^{\prime }\ and\ {r}_j\end{array}\\ {}{pZ}^{\prime}\le \kern0.5em \left(\raisebox{1ex}{$p$}\!\left/ \!\raisebox{-1ex}{$p$}\right.\right){r}_n{W_{i^{\prime }j}}^{r_n}\kern8.5em \forall {i}^{\prime }\ and\ {r}_j={r}_n\\ {}\begin{array}{c}\sum \limits_{i^{\prime }=1}^p\sum \limits_{j=1}^n{W}_{i\prime j}=1\\ {}{W}_{i\prime j}\ge 0\kern16.5em \forall {i}^{\prime }\ and\ j\end{array}\ \end{array}} $$
(5)

If we assume that Z = pZ in Model (5), Model (6) is obtained.

$$ {\displaystyle \begin{array}{c}\begin{array}{c}\mathit{\operatorname{Max}}\ \frac{Z}{p}\\ {}S.t:\\ {}Z\le \left({r}_j\left({W_{i^{\prime }j}}^{r_j}-{W_{i^{\prime }j}}^{r_j+1}\right)\right)\kern1.75em \forall {i}^{\prime } and\ {r}_j\end{array}\\ {}Z\le \kern0.5em {r}_n{W_{i^{\prime }j}}^{r_n}\kern8.75em \forall {i}^{\prime } and\ {r}_j={r}_n\\ {}\begin{array}{c}\sum \limits_{i^{\prime }=1}^p\sum \limits_{j=1}^n{W}_{i^{\prime }j}=1\ \\ {}{W}_{i^{\prime }j}\ge 0\kern12.75em \forall {i}^{\prime } and\ j\end{array}\end{array}} $$
(6)

Since p is always a positive number and there is only one variable in the objective function, solving Model (6) with \( \raisebox{1ex}{$Z$}\!\left/ \!\raisebox{-1ex}{$p$}\right. \) or Z has no effect on the optimal value of the weights. Therefore, Model (6) can be rewritten as Model (7).

$$ {\displaystyle \begin{array}{c}\begin{array}{c}\mathit{\operatorname{Max}}\ Z\\ {}S.t:\\ {}Z\le \left({r}_j\left({W_{i^{\prime }j}}^{r_j}-{W_{i^{\prime }j}}^{r_j+1}\right)\right)\kern1.75em \forall {i}^{\prime }\ and\ {r}_j\end{array}\\ {}Z\le \kern0.5em {r}_n{W_{i^{\prime }j}}^{r_n}\kern9em \forall {i}^{\prime }\ and\ {r}_j={r}_n\\ {}\begin{array}{c}\sum \limits_{i^{\prime }=1}^p\sum \limits_{j=1}^n{W}_{i\prime j}=1\\ {}{W}_{i\prime j}\ge 0\kern13em \forall {i}^{\prime }\ and\ j\end{array}\ \end{array}} $$
(7)

Hence, the optimal value of the weights in Model (4) is equal to Model (7). As a result, the optimum weights of Model (4) are equivalent to the optimum weights of Model (1).

However, it is not easy to solve a large-scale problem using the equally-weighted OPA model while there is a large number of stakeholders in the problem. Bazaraa et al. [28] addressed the complexity of the Simplex method for solving linear programming problems. They stated that for practical problems, in most cases, the iteration number is roughly between 3s/2 and 3s, where s implies the number of constraints. On the other hand, the number of constraints in the OPA method is (p × n × m) + 1, where p implies the number of experts, n implies the number of attributes, and m implies the number of alternatives. Hence, the OPA Computational Iteration (OPACI) can be estimated using Eq. (8).

$$ OPACI=\left[\frac{3p\times 3n\times 3m+3}{2},3p\times 3n\times 3m+3\ \right] $$
(8)

Since the number of alternatives is zero (m = 0) in the current study, Eq. (8) can be rewritten as Eq. (9).

$$ OPACI=\left[\frac{3p\times 3n+3}{2},3p\times 3n+3\ \right] $$
(9)

When the value of OPACI is high, the dimension of the problem should be decreased by considering the available software and hardware. Furthermore, Since the input data of the OPA are ordinal, the ordinal K-means can be employed for grouping the experts and decreasing the value of p in Eq. (9). Later, Model (10) can be used to solve the problem.

Theorem 2

Assume that the experts’ weights are equals \( \left(\raisebox{1ex}{$1$}\!\left/ \!\raisebox{-1ex}{$p$}\right.,\forall {i}^{\prime}\right) \) in a problem, and some experts provided the same opinions for the rank of all attributes. These experts can be categorized into K groups (\( {\sum}_{i=1}^K{p}_i=p \)) while the experts’ opinions in each group (pi, ∀ i) are precisely similar. In this situation, the optimum weight of Model (10) with K group of experts is equivalent to the optimum solution of Model (4) with p experts. Also, the optimum objective function of Model (10) is equal to the optimum objective function of Model (4) divided by p. As a result, the optimum weights in Model (10) are equivalent to the optimum weights of Model (1). Model (10) is called the group-weighted OPA model.

$$ {\displaystyle \begin{array}{c}\begin{array}{c}\mathit{\operatorname{Max}}\ Z\\ {}S.t:\\ {}Z\le \left(\raisebox{1ex}{$1$}\!\left/ \!\raisebox{-1ex}{${p}_i$}\right.\right)\left({r}_j\left({W_{ij}}^{r_j}-{W_{ij}}^{r_j+1}\right)\right)\kern1.75em \forall i\ and\ {r}_j\end{array}\\ {}Z\le \kern0.5em \left(\raisebox{1ex}{$1$}\!\left/ \!\raisebox{-1ex}{${p}_i$}\right.\right){r}_n{W_{ij}}^{r_n}\kern8.5em \forall i\ and\ {r}_j={r}_n\\ {}\begin{array}{c}\sum \limits_{i=1}^K\sum \limits_{j=1}^n{W}_{ij}=1\\ {}{W}_{ij}\ge 0\kern14.25em \forall i\ and\ j\end{array}\ \end{array}} $$
(10)

Proof of theorem 2

Let us assume that there are pi(∀i) experts in Model (4), who have the same opinion regarding the rank of the attributes, and we aim to categorize the experts with similar opinions in K group. Hence, the total weight of the experts with the same opinions is equal to \( {W}_{ij}={\sum}_{i\prime =1}^{p_i}{W}_{i^{\prime }j}\ \left(\forall i,{i}^{\prime } and\ j\right) \). On the other hand, since we assumed that the weights of all experts are equal, we can declare that \( {W}_{ij}={p}_i{W}_{i^{\prime }j}\ \left(\forall i,{i}^{\prime } and\ j\right) \). Moreover, instead of \( \sum \limits_{i^{\prime }=1}^p\sum \limits_{j=1}^n{W}_{i\prime j}=1 \) in Model (4), we can use \( \sum \limits_{i=1}^K\sum \limits_{j=1}^n{W}_{ij}=1 \), which are equivalent. Considering the aforementioned information, Model (4) can be rewritten as Model (11).

$$ {\displaystyle \begin{array}{c}\mathit{\operatorname{Max}}\ {Z}^{\prime}\\ {}S.t:\\ {}\begin{array}{c}{Z}^{\prime}\le \left(\raisebox{1ex}{$1$}\!\left/ \!\raisebox{-1ex}{$p$}\right.\right)\left({r}_j\left({W}_{ij}/{p_i}^{r_j}-{W}_{ij}/{p_i}^{r_j+1}\right)\right)\kern1.75em \forall i\ and\ {r}_j\\ {}{Z}^{\prime}\le \kern0.5em \left(\raisebox{1ex}{$1$}\!\left/ \!\raisebox{-1ex}{$p$}\right.\right){r}_n{W}_{ij}/{p_i}^{r_n}\kern9em \forall i\ and\ {r}_j={r}_n\\ {}\begin{array}{c}\sum \limits_{i=1}^K\sum \limits_{j=1}^n{W}_{ij}=1\ \\ {}{W}_{ij}\ge 0\kern17em \forall i\ and\ j\end{array}\end{array}\end{array}} $$
(11)

If we multiply Model (11) by p, which is a positive number, Model (12) is obtained:

$$ {\displaystyle \begin{array}{c}\mathit{\operatorname{Max}}\ {Z}^{\prime}\\ {}S.t:\\ {}\begin{array}{c}p{Z}^{\prime}\le \left({r}_j\left({\frac{W_{ij}}{p_i}}^{r_j}-{\frac{W_{ij}}{p_i}}^{r_j+1}\right)\right)\kern1.75em \forall i\ and\ {r}_j\\ {}p{Z}^{\prime}\le \kern0.5em {r}_n{\frac{W_{ij}}{p_i}}^{r_n}\kern10em \forall i\ and\ {r}_j={r}_n\\ {}\begin{array}{c}\sum \limits_{i=1}^K\sum \limits_{j=1}^n{W}_{ij}=1\ \\ {}{W}_{ij}\ge 0\kern14.75em \forall i\ and\ j\end{array}\end{array}\end{array}} $$
(12)

Let us assume that Z = pZ in Model (12). Therefore, Model (12) can be rewritten as Model (13).

$$ {\displaystyle \begin{array}{c}\begin{array}{c}\mathit{\operatorname{Max}}\ Z/p\\ {}S.t:\\ {}Z\le \left({r}_j\left({W}_{ij}/{p_i}^{r_j}-{W}_{ij}/{p_i}^{r_j+1}\right)\right)\kern1.75em \forall i\ and\ {r}_j\end{array}\\ {}Z\le \kern0.5em {r}_n{W}_{ij}/{p_i}^{r_n}\kern10em \forall i\ and\ {r}_j={r}_n\\ {}\begin{array}{c}\sum \limits_{i=1}^K\sum \limits_{j=1}^n{W}_{ij}=1\ \\ {}{W}_{ij}\ge 0\kern13.25em \forall i\ and\ j\end{array}\end{array}} $$
(13)

Since p is always a positive number and there is one variable in the objective function, solving Model (13) with \( \raisebox{1ex}{$Z$}\!\left/ \!\raisebox{-1ex}{$p$}\right. \) or Z has no effect on the optimal value of the weights. Therefore, Model (13) can be rewritten as Model (14).

$$ {\displaystyle \begin{array}{c}\begin{array}{c}\mathit{\operatorname{Max}}\ Z\\ {}S.t:\\ {}Z\le \left(\raisebox{1ex}{$1$}\!\left/ \!\raisebox{-1ex}{${p}_i$}\right.\right)\left({r}_j\left({W_{i\mathrm{j}}}^{r_j}-{W_{ij}}^{r_j+1}\right)\right)\kern1.75em \forall i\ and\ {r}_j\end{array}\\ {}Z\le \kern0.5em \left(\raisebox{1ex}{$1$}\!\left/ \!\raisebox{-1ex}{${p}_i$}\right.\right){r}_n{W_{ij}}^{r_n}\kern8.5em \forall i\ and\ {r}_j={r}_n\\ {}\begin{array}{c}\sum \limits_{i=1}^K\sum \limits_{j=1}^n{W}_{ij}=1\\ {}{W}_{ij}\ge 0\kern13.25em \forall i\ and\ j\end{array}\ \end{array}} $$
(14)

As a result, the optimal values of the weights in Model (4) are equal to the optimal weights in Model (10).

An application of theorem 2

The proposed Model (10) can consider the priority of the clusters on the weights of the attributes using the number of members in each cluster. The application of Model (10) is not limited to this study and may be used to consider the years of work experience of the experts instead of ordinal preferences while calculating the weights of attributes. For this purpose, years of experience of expert i can be considered as pi in Model (10).

3.2 Ordinal K-means

Clustering has a broad application in real-world problems. Especially when there is a large number of data, it is challenging to analyze them due to computational time and cost. In such a situation, clustering can contribute to the decision-maker to group the data based on one or more distinguishing factors. Cheng and Leu [29] addressed the application of clustering in construction management. In another study, Pham and Afify [30] discussed applying the clustering methods in engineering. Among clustering methods, K-means is one of the most well-known methods [31, 32]. For the first time, K-means was presented by Lloyd in 1982 [33]. K-means was used widely for solving LSGDM problems as well. Since the input of the current study are ordinal data, we aim to employ the ordinal K-means, which was proposed by Tang et al. [34]. However, this algorithm is adopted to be used in the OPA model. The steps of the ordinal K-means are presented as follows:

  • Step 1: by starting the algorithm, several initializations should be provided. Also, the iteration number should be placed one (I = 1).

  • Step 2: In this step, initial clustering points should be selected randomly.

  • Step 3: Using Euclidean distance, the distance from each point to each center should be computed. Eq. (15) can determine the distance from each point to each center.

$$ {d}_{Euc}=\sqrt{\frac{\sum \limits_{j=1}^n{\left({r_j}^{\left({i}^{\prime}\right)}-{r_j}^{(C)}\right)}^2}{2\left[{\left(n-1\right)}^2+{\left(n-3\right)}^2+\dots \right]}} $$
(15)

It should be noted that when the value of n is even, the last sentence of (n − 1)2 + (n − 3)2 + … is 1. When the value of n is odd, the last sentence of (n − 1)2 + (n − 3)2 + … is 2. It should be noted that dEuc ∈ [0, 1] because (n − 1)2 + (n − 3)2 + … is the biggest quadratic difference between two rank series.

  • Step 4: In this step, each point should be assigned to the closest cluster center based on the obtained value in Eq. (15).

  • Step 5: In this step, each cluster center should be updated. For calculating the cluster center, the preference vector ought to be used. Assume that \( {\omega}^{\left({i}^{\prime}\right)}={\left({\omega}_1^{\left({i}^{\prime}\right)},{\omega}_2^{\left({i}^{\prime}\right)},\dots, {\omega}_n^{\left({i}^{\prime}\right)}\right)}^T \)is the preference vector regarding the attributes by expert i. The preference vector should be calculated using Eq. (16)

$$ {\omega}_j^{\left({i}^{\prime}\right)}=\left(n-{r_j}^{\left({i}^{\prime}\right)}\right)/\sum \limits_{j=1}^{n-1}j $$
(16)

After that, the clustering center for cluster Ci can be computed utilizing Eq. (17)

$$ {\omega}_j^{\left({C}_i\right)}=\frac{1}{p_i}\sum \limits_{i=1}^{p_i}{\omega}_j^{\left(i{i}^{\prime}\right)} $$
(17)

Where \( {\omega}^{\left(i{i}^{\prime}\right)}={\left({\omega}_1^{\left(i{i}^{\prime}\right)},{\omega}_2^{\left({ii}^{\prime}\right)},\dots, {\omega}_n^{\left(i{i}^{\prime}\right)}\right)}^T \) implies the preference vector of the i′th expert in the cluster Ci, and pi implies the number of experts in cluster i. In this study, the weights of all experts are considered equally. In order to consider the priority of the clusters based on the number of experts, the value of \( {r}_{i^{\prime }} \) in the OPA model can be calculated using Eq. (18), which is shown in Model (10) as well.

$$ {r}_{i^{\prime }}=\frac{1}{p_i} $$
(18)

Where pi implies the number of experts in cluster i.

3.3 Silhouette coefficient

Rousseeuw [35] proposed a graphical method to check the quality of the clustering, namely the Silhouette plot. The Silhouette plot was established based on the Silhouette coefficient value. The silhouette coefficient is a number between −1 and + 1 constantly. The value of +1 shows a strong association with a cluster as well as the distance from neighbouring clusters. The value of 0 implies no distinct association, and the value of −1 indicates the wrong assignment to a cluster [36]. Silhouette coefficient is employed by scholars widely to calculate the optimal number of clusters [37, 38]. Let us assume that A and C are two different clusters and i ∈ A. The value of the Silhouette coefficient can be calculated utilizing Eq. (19) or Eq. (20).

$$ s\left({i}^{\prime}\right)=\frac{b\left({i}^{\prime}\right)-a\left({i}^{\prime}\right)}{\mathit{\max}\left\{a\left({i}^{\prime}\right),b\left({i}^{\prime}\right)\right\}} $$
(19)

or

$$ s\left({i}^{\prime}\right)=\left\{\begin{array}{c}1-\frac{a\left({i}^{\prime}\right)}{b\left({i}^{\prime}\right)}\kern2.5em if\ a\left({i}^{\prime}\right)<b\left({i}^{\prime}\right)\\ {}0\kern6.5em if\ a\left({i}^{\prime}\right)=b\left({i}^{\prime}\right)\\ {}\frac{b\left({i}^{\prime}\right)}{a\left({i}^{\prime}\right)}-1\kern3em if\ a\left({i}^{\prime}\right)>b\left({i}^{\prime}\right)\ \end{array}\right. $$
(20)

Where

a(i):

= Average dissimilarity of expert i to all other experts in cluster A.

b(i):

= MinmumC ≠ A dEuc(i, C)

Therefore, a higher value of the Silhouette coefficient indicates better quality of clustering. It is essential to check the Silhouette plot during clustering analysis to decrease the error of false clustering. This issue is addressed in the current study carefully because low-quality clustering may change the solution of the problem entirely.

3.4 Confidence level measurement

The consensus among the experts plays a vital role in working with ordinal data. The degree of association among the experts’ opinions should be significant which shows the reliability of the input data. It should be noted that Kendall’s W is a non-parametric statistic test that can determine the concordance among the ranks which are provided by various experts [39, 40]. However, this method cannot consider the weight of the experts for calculating the degree of association among experts’ opinions. Hence, we aim to propose Weighted Kendall’s W (Ww) and employ it for calculating the consensus in each cluster to check the reliability of the input data. The MATLAB source file of the proposed approach is available on https://mathworks.com/matlabcentral/fileexchange/116775-weighted-kendall-s-w. The steps of the Weighted Kendall’s W are presented as follows:

Assume that i (1, 2, …, K) represents the index of the clusters, and j implies the index of attributes. rij implies the rank of attribute j by cluster i. ϑi implies the weight of the cluster i (\( {\sum}_{i=1}^K{\vartheta}_i=1 \)), which is \( \raisebox{1ex}{${p}_i$}\!\left/ \!\raisebox{-1ex}{$p$}\right. \) in the current study. The total rank of attribute j (j = 1, 2, …, n) by all experts in cluster i can be calculated using Eq. (21).

$$ {R}_{ij}=\sum \limits_{i=1}^K{\vartheta}_i\times {r}_{ij}\kern2.25em \forall j $$
(21)

After that, mean value of the attributes’ ranks should be calculated by utilizing Eq. (22)

$$ \overline{R}=\frac{1}{n}\sum \limits_{j=1}^n{R}_{ij} $$
(22)

Eq. (23) determines sum of squares of the observed deviations from the mean \( \overline{R} \).

$$ S=\sum \limits_{j=1}^n{\left({R}_{ij}-\overline{R}\right)}^2 $$
(23)

Using the obtained value in Eq. (23), the value of Weighted Kendall’s W can be calculated with the aid of Eq. (24)

$$ {W}_w=\frac{12S}{\left({n}^3-n\right)}\kern1.25em $$
(24)

Original Kendall’s W is a special case of Weighted Kendall’s W. When ϑi has a uniform distribution, the value of Kendall’s W and Weighted Kendall’s W are equal.

Theorem 3

Kendall’s W is a special case of Weighted Kendall’s W. The value of Kendall’s W is equal to Weighted Kendall’s W when the distribution of the weights is uniform.

Proof of theorem 3

Assume that there are K clusters (raters) in the problem. If we assume that the weights of the clusters are equal, the weight of each cluster is \( \frac{1}{K} \), where the summation of the weights is equal to 1. After replacing this value in Eq. (21), the revised formula of Rij is as follows:

$$ {R_{ij}}^{\prime }=\frac{1}{K}\sum \limits_{i=1}^K{r}_{ij}=\frac{R_{ij}}{K} $$

Considering Eq. (22), the revised formula of \( \overline{R} \) is resulted as follows:

$$ {\overline{R}}^{\prime }=\frac{1}{nK}\sum \limits_{j=1}^n{R}_{ij}=\frac{\overline{R}}{K} $$

Later, the revised formula of S is obtained as follows using Eq. (23):

$$ {S}^{\prime }=\sum \limits_{j=1}^n{\left(\frac{R_{ij}}{K}-\frac{\overline{R}}{K}\right)}^2=\frac{1}{K^2}\sum \limits_{j=1}^n{\left({R}_{ij}-\overline{R}\right)}^2=\frac{S}{K^2} $$

Based on Eq. (24), the revised formula of Weighted Kendall’s W is as follows:

$$ {W_w}^{\prime }=\frac{12S}{K^2\left({n}^3-n\right)} $$

As we can see, the above formula is equal to the original Kendall’s W. Therefore, Kendall’s W is equal to Weighted Kendall’s W when the distribution of the weights is uniform, and Theorem 3 is proved.

When there is a tie in the rank of the attributes, Eq. (24) cannot be used. Also, instead of tie ranks, the mean value of the positions should be used. In this situation, Eq. (25) should be utilized to determine the number of ties.

$$ {T}_i=\sum \limits_{j=1}^n\left({t}_{ij}^3-{t}_{ij}\right)\kern2.75em \forall i $$
(25)

tij represents the number of tie ranks in cluster i and attribute j. Ti implies the number of ties in cluster i. After that, the value of the corrected Weighted Kendall’s W should be determined using Eq. (26)

$$ {W}_w=\frac{12S}{\left({n}^3-n\right)-{\sum}_{i=1}^K{\vartheta}_i{T}_i} $$
(26)

Theorem 4

Corrected Kendall’s W is a special case of corrected Weighted Kendall’s W. The value of corrected Kendall’s W is equal to corrected Weighted Kendall’s W when the distribution of the weights is uniform.

Proof of theorem 4

From Theorem 3, we are informed that,

$$ {S}^{\prime }=\frac{S}{K^2} $$

Also, the weight of each cluster is \( \frac{1}{K} \) . If we replace these values in Eq. (26), the revised formula of corrected Weighted Kendall’s W is as follows:

$$ {W_w}^{\prime }=\frac{12\frac{S}{K^2}}{\left({n}^3-n\right)-{\sum}_{i=1}^K\frac{1}{K}{T}_i}=\frac{12S}{K^2\left({n}^3-n\right)-K{\sum}_{i=1}^K{T}_i} $$

As we can see, the above formula is equal to the original corrected Kendall’s W. Therefore, corrected Kendall’s W is equal to corrected Weighted Kendall’s W when the distribution of the weights is uniform and Theorem 4 is proved.

Mahmoudi and Javed [41] proposed a confidence level index for the OPA model. However, they used the original Kendall’s W in their study based on the need. In this study, we extended their index by employing the proposed Weighted Kendall’s W to consider the weights of the cluster during confidence level measurement. For this purpose, the x statistic should be calculated by using Eq. (27) with the degrees of freedom v1 and v2.

$$ x=\frac{W_w\left(K-1\right)}{1-{W}_w}\kern3.75em {v}_1=n-1-\left(2/K\right)\kern0.5em and\ {v}_2=\left(K-1\right){v}_1 $$
(27)

Based on F-distribution, the value of the probability density function of x can be determined using Eq. (28) [42].

$$ f\left(x;{v}_1,{v}_2\right)=\frac{\Gamma \left(\frac{v1+v2}{2}\right){\left(\frac{v_1}{v_2}\right)}^{\frac{v_1}{2}}{x}^{\frac{v_1}{2}-1}}{\Gamma \left(\frac{v1}{2}\right)\ \Gamma \left(\frac{v2}{2}\right)\ {\left(1+\frac{v_1}{v_2}x\right)}^{\frac{v_1+{v}_2}{2}}}\kern1em x>0 $$
(28)

While gamma function (Γ) is defined in Eq. (29).

$$ \Gamma (a)={\int}_0^{\infty }{t}^{a-1}{e}^{-t} dt $$
(29)

Finally, guided by the cumulative distribution function (CDF), the value of the confidence level should be determined by employing Eq. (30).

$$ CL=F\left(x;{v}_1,{v}_2\right)={\int}_0^xf\left(x;{v}_1,{v}_2\right) dx $$
(30)

The value of the confidence level in 0.99, 0.95, 0.90 shows the significant test in levels 0.01, 0.05, 0.1 by F-distribution. Therefore, these points can be utilized as the threshold for the confidence level. Hence, the thresholds can be defined as in Fig. 1.

Fig. 1
figure 1

The objective threshold for the confidence level [41]

4 The proposed framework

In this section, the steps of the proposed framework are explained. There are three main steps to calculate the performance of the healthcare construction projects, which are shown in Fig. 2. The first step focuses on the data collection and defining the problem, Step 2 tries to decrease the size of the problem and Step 3 measures the performance of the projects.

Fig. 2
figure 2

The steps of the proposed approach

  • Step 1: First of all, the indicators for performance measurement must be defined. Selecting the indicators depends on several factors, including the availability of the data, feasibility, importance, etc. After that, we should collect the experts’ opinions associated with the indicators. The opinions should be in the form of ordinal data. On the other hand, the quantitative data related to the projects in the indicators should be collected, e.g., delay of the projects, cost of construction, etc.

  • Step 2: In this step, the complexity of the problem should be checked first. The complexity can be represented by the number of constraints or computational iteration formula in Eq. (9). The user ought to consider available resources (e.g., software, hardware, and effort of solving) before solving the problem. For example, if the user would like to use MS Excel software, he/she should be informed that the Excel solver cannot handle a large number of constraints. Hence, the acceptable range of complexity should be defined by the user. Next, a Silhouette plot should be provided using any software such as MATLAB within the acceptable range of complexity (See Eq. (20)). If there is any value higher than 0.5 for the Silhouette coefficient, we can select the number of clusters with the highest Silhouette coefficient value and run the ordinal K-means algorithm. Otherwise, a feedback mechanism should be applied, which includes various actions such as adding more experts, discussing with the experts to make their opinions more consistent, etc. Later, the value of the confidence level should be determined using Eqs. (21) to (30), and the obtained value must be compared with the thresholds based on the sensitivity level of the system. If it can meet the required threshold, we can go to the next step. Otherwise, the feedback mechanism must be applied, and Step 2 should be repeated.

  • Step 3: With making the final decision regarding the number of clusters, the conventional problem can be solved by employing Model (10). Next, the weights of the clusters and the weights of the indicators are obtained. Indeed, the LSGDM is employed for calculating the weight of the indicators for performance measurement. Finally, the relative project performance (RPP) can be determined using Eq. (31) for each project.

$$ \mathrm{RPP}=\sum \limits_{j=1}^n{W}_j\times {\hat{x}}_{jh}\kern2em \forall h\kern1.25em h=1,2,\dots .,q $$
(31)

Where Wj is the weight of indicator j and \( {\hat{x}}_{jh} \) is the normalized value of the quantitative data of project h in indicator j. When the indicator is benefit-type, Eq. (32) should be employed, and when the indicator is cost-type, Eq. (33) should be utilized to normalize the data in each indicator [43].

$$ {{\hat{x}}_{jh}}^{+}=\frac{x_{jh}-{\mathit{\min}}_j\ \left({x}_{jh}\right)}{{\mathit{\max}}_j\ \left({x}_{jh}\right)-{\mathit{\min}}_j\ \left({x}_{jh}\right)}\kern5em \forall j $$
(32)
$$ {{\hat{x}}_{jh}}^{-}=\frac{{\mathit{\max}}_j\left({x}_{jh}\right)-{x}_{jh}}{{\mathit{\max}}_j\ \left({x}_{jh}\right)-{\mathit{\min}}_j\ \left({x}_{jh}\right)}\kern4.5em \forall j $$
(33)

After determining the RPP of the healthcare construction projects, they can be compared, and the strengths and weaknesses of the projects in each indicator can be investigated.

5 Comparative analysis

Tang et al. [34] presented an illustrative example related to the public transportation system in China. They have provided the input data as Table 2, which are collected from 20 experts for 5 alternatives. In this section, the proposed approach is compared with the approach proposed by Tang et al. [34] to check the reliability and feasibility of both frameworks.

Table 2 Ordinal preference for the illustrative example

To illustrate the complexity of the problem, the value of the OPACI is determined using Eq. (9) and depicted in Fig. 3 for a various number of clusters. Here, we assume that considering the available software, hardware, and time, the acceptable range of complexity is between i = 2 to i = 6. It can be different for various users based on their situations.

Fig. 3
figure 3

The complexity of the illustrative example

Therefore, we should draw the Silhouette plot for the aforementioned range of complexity. Using the MATLAB software, the Silhouette plot is illustrated in Fig. 4. As can be seen from Fig. 4, i = 3 reached the highest Silhouette coefficient value among others, with a value higher than 0.5. Hence, we select i = 3 and run the ordinal K-means.

Fig. 4
figure 4

The Silhouette plot for the presented data in Table 2

After applying the ordinal K-means to the data in Table 2, the cluster numbers are shown in the last column of Table 2 and the cluster centers are depicted in Table 3.

Table 3 Cluster centers after applying ordinal K-means on the data in Table 2

Since there is a tie rank in cluster number 1, we should replace the mean value instead of them (the mean value of the second and third positions is 2.5). Therefore, for cluster number 1, we have 4, 1, 2.5, 5, 2.5 respectively, to calculate the Weighted Kendall’s W. If we calculate the confidence level for the cluster centers in Table 3 using MATLAB software, its value is 0.7332, which is truly low and cannot meet the required threshold (CL = 0.90 equals a significant level of 0.1). Therefore, the feedback mechanism should be done to improve the consistency ratio of the input data. Also, Tang et al. [34] faced the same scenario, and they mentioned that the global consensus index value could not meet the threshold. After applying the feedback mechanism, Tang et al. [34] presented the data in Table 4.

Table 4 The input data after applying the feedback mechanism

Employing the MATLAB software, the Silhouette plot is provided in Fig. 5, which shows that the highest value of the Silhouette coefficient is associated with i = 3.

Fig. 5
figure 5

The Silhouette plot for the presented data in Table 4

After applying the ordinal K-means to the data in Table 4, the cluster numbers are shown in the last column of Table 4 and the cluster centers are depicted in Table 5.

Table 5 Cluster centers after applying ordinal K-means on the data in Table 4

If we calculate the confidence level for the cluster centers in Table 5 using MATLAB software, its value is 0.9182, which is suitable for sensitive problems and can meet the required threshold (CL = 0.90 equals the significant level of 0.1). It is worth mentioning that Tang et al. [34] calculated the consensus index in their study using their proposed index. They also confirmed that the data in Table 4 meets the threshold.

After that, we can solve the problem based on the data in Table 5 and Model (10) to calculate the weights of alternatives. The weights are resulted in Table 6 and compared with those which were obtained by Tang et al. [34].

Table 6 The comparative analysis

In order to compare the ranks in Table 6, the Spearman’s rank-order correlation is utilized, which is a non-parametric test, and its formula is presented in Eq. (34) [44, 45].

$$ {r}_{\mathrm{rho}}=1-\left[\frac{6.{\sum}_{j=1}^n{\left({d}_j\right)}^2}{n.\left({n}^2-1\right)}\right] $$
(34)

where dj is the difference between two ranks, and n is the number of alternatives. Based on Table 6, the Spearman’s rank-order correlation is calculated using IBM SPSS software (version 26), and the results are presented in Table 7.

Table 7 The correlation between the results

As can be seen from Table 7, there is a strong correlation between the obtained ranks by the proposed approach and Tang et al. [34].

6 Case study

The policymakers in China look forward to finding a reasonable method for measuring the performance of the healthcare construction projects. It is very important to find out which projects are successful and identify the reasons to consider in future healthcare projects. As mentioned earlier, measuring the performance of healthcare projects is a challenging task, and it should be established based on all stakeholders’ opinions. Since there are a lot of stakeholders involved in each healthcare construction project, it is not easy to solve such large-scale problems. In the current study, five healthcare construction projects from China are selected for performance measurement and evaluation. The selected hospitals are located in Nanjing, Beijing, Guangzhou, Quzhou, and Dalian. The construction phase of these projects has been finished already.

The quantitative information associated with the projects in each indicator is presented in Table 8. Here, there are five indicators for project performance measurement, which can be more or less based on the need.

Table 8 The quantitative information regarding each project in each indicator for the case study

In a survey, we collected the experts’ opinions related to performance indicators in Table 8. 43 experts participated in the survey. The experts’ opinions associated with the indicators are presented in Table 9.

Table 9 The experts’ opinions regarding the indicators for the case study

The value of OPACI for the various number of clusters is illustrated in Fig. 6. As shown in Fig. 6, the problem for i = 43 requires lots of iterations to be solved. Considering the available software and hardware, the acceptable range of complexity is between i = 2 to i = 9 for this case study.

Fig. 6
figure 6

The complexity of the problem for various cluster numbers in the case study

Then, the Silhouette plot is provided for the acceptable complexity range which is illustrated in Fig. 7. As can be seen from Fig. 7, the highest value of the Silhouette coefficient is associated with i = 9. Hence, it can be selected as the optimal number of the clusters.

Fig. 7
figure 7

The Silhouette plot for the case study

We applied the ordinal K-means to the data in Table 9, the cluster numbers are shown in the last column of Table 9. The centers of the clusters for i = 9 are illustrated in Table 10. Moreover, the number of experts in each cluster is provided, which is necessary to solve the problem.

Table 10 Cluster center for various cluster numbers

Here, we calculated the confidence level for that data in Table 10 employing Eqs. (21) to (30). The confidence level is reported as 0.9523. Based on the thresholds in Fig. 1, the value of the confidence level is suitable for very sensitive problems. Hence, we do not need to apply a feedback mechanism and can go to the next step. Considering the data in Table 10 and Model (10), the weight of the attributes is obtained in Table 11.

Table 11 The weight and rank of the indicators through Model (10)

Finally, using Tables 8 and 11, the relative performance of the hospital projects has resulted in Table 12.

Table 12 The relative performance of the hospital projects

Based on Table 12, the project of Hospital 5 achieved the highest relative performance with a value of 0.65109. After that, Hospital 1 achieved the second position. These projects can be a perfect benchmark for implementing future projects, and the reason for their suitable performance can be extracted and considered as a lesson learned.

7 Sensitivity analysis

For more investigation, we have provided a comprehensive sensitivity analysis of the case study. First, the confidence level is calculated for a various number of clusters. As can be seen from Fig. 8, the confidence level is experiencing an increasing trend. There are two points which met the confidence level threshold, including i = 8 and i = 9. This graph shows that i = 9 is the best one to be selected as the optimal number of clusters with the highest confidence level value.

Fig. 8
figure 8

The sensitivity of the confidence level for the case study

Also, we have calculated the weights of the attributes for various number of clusters to check the sensitivity of the weights. The results are shown in Table 13 in detail and compared in Fig. 9. As can be seen from Fig. 9, the error for i = 2, i = 3, and i = 4 is significant, which shows that they are not appropriate for selecting as the optimal number of clusters. On the other hand, the weights of the attributes for i = 9 and i = 43 are equivalent. The main reason is Silhouette coefficient value equals 1, which is shown in Fig. 7. It shows that the data are clustered perfectly, and the quality of clustering is significantly high. However, the error among i = 6, i = 7, i = 8, i = 9 is not significant, which shows that all of them are correlated.

Table 13 The sensitivity of the indicators’ weights for the case study
Fig. 9
figure 9

The sensitivity of the indicators’ weights for the case study

The relative performance of the hospitals is calculated for a various number of clusters which are shown in Table 14 and Fig. 10. As we expected, the relative performance of the hospitals is precisely the same for i = 9 and i = 43. Also, Hospital 5 is the best option in all conditions with a slight change for the various number of clusters.

Table 14 The relative performance of the hospital projects for various cluster numbers
Fig. 10
figure 10

The relative performance of the hospital projects for various cluster numbers

Based on the aforementioned sensitivity analysis, the proposed framework works logically and can decrease the volume of the calculations efficiently. Indeed, we solved a small-scale problem (i = 9) instead of a large-scale problem (i = 43) while the results are entirely the same in both conditions. Thus, one can argue that the proposed approach can be used confidently for LSGDM problems while saving resources (efforts, time, etc.).

8 Discussion and conclusion

Performance measurement of healthcare construction projects plays a vital role in the success of future healthcare projects. The weaknesses and strengths should be identified and considered as a lesson learned for future projects. To determine the performance of the construction healthcare projects, there is a need to define several performance indicators. The degree of importance of the indicators should be specified by a large number of stakeholders, which can lead to more reasonable results, and finally, with the aid of the quantitative data regarding each project in each indicator, the performance of the projects can be determined. However, considering the opinions of a large number of stakeholders is not an easy job, and it requires a significant volume of calculations. To overcome this obstruction, a novel framework was proposed that works using preference data, and it was compared with the system proposed by Tang et al. [34].

Based on the comparative analysis, one can see several benefits in the proposed framework. The first one is related to the computational complexity during the selection of the optimal number of clusters. Indeed, one of the inputs for the illustration of the Silhouette plot is the range of the possible number of clusters. This study considers this matter before drawing the Silhouette plot, which was ignored in the study of Tang et al. [34]. The second benefit of the proposed framework is providing a confidence level index which considers various objective threshold levels based on the sensitivity of the problem. It will increase the flexibility of the proposed framework in the real-world situations, which was ignored by other scholars. For example, Tang et al. [34] offered one type of threshold for any problem. The third benefit of the proposed framework is decreasing the cost of calculation by running the feedback mechanism as earlier as possible. In other words, it is not essential to solve the problem till the end and obtain the final rank to understand whether there is a need for a feedback mechanism. There are various layers in the proposed framework to check the need for the feedback mechanism. It can decrease unnecessary computations and optimize the algorithm.

This study can be extended to solve LSGDM problems, including experts, attributes, and alternatives simultaneously using the OPA in the future. Indeed, the current framework can determine only the weights of the attributes in LSGDM problems. Moreover, metaheuristic algorithms such as particle swarm optimization (PSO) can be used to extract the optimal number of clusters that may lead to more accurate results. The current study can be extended to handle the uncertainty in preference data, such as incomplete input data, because the experts may not be sure about their opinions in real-world situations.