Introduction

With the rapid development of society and technology, decision problem becomes increasingly complicated which contains usually much uncertain information. A decision problem may involve various fields. For instance, when a company makes a selection of investors, it will consider many factors of investor such as financial capacity, credit risks, industry familiarity and so on. In traditional group decision making, a small number of experts are hard to give comprehensive evaluations for schemes. Thus, large-scale group decision making (LSGDM) is introduced, in which a large number of decision makers (DMs) who come from distinct fields participate the process of decision making to ensure accuracy and efficiency of decision results. Normally, there are more than 20 (contain 20) DMs in LSGDM. Large-scale group decision making has attracted much attention of researchers [1, 2] and it is applied to many aspects, e.g., high-speed rail system [3], earthquake shelter selection [4], and healthcare service [5].

Due to the cognitive complexity of human, when conducting decision making, it is more appropriate to use fuzzy relations to depict people’s evaluations [6]. Besides, compared with numeral rating, DMs tend to adopt linguistic labels because of its directness and simplicity. In fuzzy decision making, Zadeh [7] introduced the concept of fuzzy sets (FSs) and Torra [8] proposed hesitant fuzzy sets (HFSs) which well express the hesitancy and fuzziness of DMs. Cui et al. [9] utilized the HFSs to measure product similarity. Based on linguistic variables and HFSs, Rodriguez et al. [10] developed the theory of hesitant fuzzy linguistic term sets (HFLTSs) so that DMs can use a linguistic term set (LTS) to describe their preferences. On the basis of HFLTS, Chen et al. [11] proposed the proportional HFLTSs (PHFLTSs) and explored the aggregation of HFLTS possibility based on K-means clustering [12]. Nevertheless, the defect of HFLTSs is that the weights or importance of terms are viewed as identical ones. To overcome the shortcoming, Pang et al. [13] developed the concept of probabilistic linguistic term sets (PLTSs) which reflect flexibly ambiguous information of DMs and importance distribution of LTs. PLTS is an useful tool in decision making, especially in LSGDM, because experts can utilize it to give more accurate and actual evaluations. In multi-attribute LSGDM problem, the decision matrix with PLTS is often used to represent evaluation information of DMs with respect to alternatives and attributes. However, in most cases, owing to limited knowledge and time urgency, experts are not able to provide a complete decision matrix. Thus, before conducting decision making, the incomplete matrix should be filled.

Many scholars have proposed various methods to estimate the missing information in the incomplete matrix and we have a brief review below. From the respective of estimation strategies, Ureña et al. [14] summarized as individual strategy and social strategy that the former only considers information of incomplete decision matrix while the latter needs to refer to the decision information of other experts. Based on additive consistency and intermediate value, Herrera-Viedma et al. [15] proposed an iterative process to estimate the missing preference values, which avoids the contradiction among preferences. Some researchers filled incomplete decision information according to consistency level of group [16]. Meng et al. [17] utilized the goal programming model to estimate missing values of decision matrix under fuzzy preference relation. Capuano et al. [18] developed an iteration process to solve the problem of incomplete evaluation information based on additive consistency and transitivity. Guo et al. [19] proposed the least deviation model to estimate missing preference, which achieved the maximum confidence under the mechanism of consistency and the range of missing values. Liang et al. [20] estimated incomplete information by means of collaborative filtering algorithm. Although various methods have been proposed to estimate incomplete decision matrix, few researches take the cognitive level of experts into account, which is an important factor to ensure the accuracy of estimated values. When estimating the missing information, we usually refer to the related information of other experts. If the referenced experts have low cognitive level, the accuracy of their evaluation information will be also poor. Thus, to solve this problem, a novel estimation method is proposed based on knowledge-math degree with reliability in this paper.

To apply the proposed method to LSGDM, we develop a decision process including estimating incomplete information, subgroup detecting, consensus reaching process (CRP) and selection process. With respect to subgroup detecting, some methods are proposed such as fuzzy C-means algorithm [23], K-means method [3], the approach based on supporting degree of alternatives [24] and variable grouping method [25]. In this paper, due to the efficiency and superiority of K-means, we adopt K-means method to cluster the subgroups. Besides, consensus reaching process has been the research focus in LSGDM and many CRP models are developed [26,27,28], e.g., two-stage CRP model [21], the CRP model based on conflict degree [22]. To intuitively analysis previous researches, the Table 1 is given. Due to the large amount of calculation in previous CRP of LSGDM, we construct a new CRP based on social trust relationship, which can also remain the original evaluation information to some extent. The main contributions in this paper are as follows.

  1. (1)

    The new concepts of knowledge-match degree and PLTS reliability are introduced to measure the importance of referenced experts. On the basis of knowledge-match degree, we integrate the reliability into knowledge-match degree, namely knowledge-match degree with reliability, which is used to measure the cognitive level of experts.

  2. (2)

    Based on knowledge-match degree with reliability, a novel method is developed to estimate the missing information in incomplete matrix represented by PLTS. To apply the method better, we design a algorithm to realize it.

  3. (3)

    To solve the LSGDM problem, we construct a LSGDM process in term of the proposed estimation method. In the LSGDM process, a new CRP is constructed based on trust relationship.

Table 1 Literature review

The rest contents are arranged as follows. In the next section, some basic knowledge related to PLTS are reviewed including definitions, operations, normalization and so on. Subsequently, we propose the novel method to estimate incomplete decision matrix with PLTS based on knowledge-match degree with reliability in the third section. In the fourth section, the LSGDM process is developed based on the proposed estimation method and a new CRP is construct based on trust relationship. A case about pharmaceutical manufacturer selection is used to illustrate the proposed decision method in the fifth section. To verify effectiveness and superiority of the proposed methods, we make the comparative analysis with other methods in the sixth section, and finally, draw a conclusion in the last section.

Preliminary

In this section, some basics about PLTS are introduced to prepare for latter theoretical method. The PLTS is proposed by Pang [13], and the comparison rule, the normalized method, and some operations are also given.

PLTS

Definition 1

[13] Suppose that S is a linguistic term set (LTS), S = {sα = 0,1,2,…,2τ}. Let L(p) be a PLTS:

$$ \begin{aligned} L(p)& = \left\{ L^{k} (p^{k} ){|}L^{k} \in S,p^{k} \ge 0,k = 1,2,...,\# L(p),\right.\\ & \quad \left.\times \sum\limits_{k = 1}^{\# L(p)} {p^{k} \le 1} \right\} ,\end{aligned} $$

where Lk and pk denote the kth term and its probability in L(p). #L(p) is the number of terms in L(p). If \(\sum\nolimits_{k = 1}^{\# L(p)} {p^{k} < 1}\), the PLTS is incomplete, which should be normalized before calculation or aggregation. According to Pang’s method [13], the normalized PLTS is given by

$$ \dot{L}(p) = \{ L^{k} (\dot{p}^{k} )|k = 1,2,...,\# L(p)\} , $$
(1)

where \(\dot{p}^{k} = p^{k} /\sum\nolimits_{k = 1}^{\# L(p)} {p^{k} }\).

Besides, when calculate two PLTSs, the cardinality of terms should be uniformed. Given any two PLTSs, L1(p) and L2(p), if #L1(p) > #L2(p), add #L1(p)-#L2(p) terms to L2(p) and the added terms are the least ones in L2(p). The probabilities of added elements are set 0. If #L2(p) > #L1(p), it is the same as above.

Pang [13] also proposed some operations of PLTS, which are listed as follows”.

  1. (1)

    (1)\(L_{1} (p) \oplus L_{2} (p) = \cup_{{L_{1}^{k} \in L_{1} (p),L_{2}^{k} \in L_{2} (p)}} \{ p_{1}^{k} L_{1}^{k} \oplus p_{2}^{k} L_{2}^{k} \}\).

  2. (2)

    \(L_{1} (p) \otimes L_{2} (p) = \cup_{{L_{1}^{k} \in L_{1} (p),L_{2}^{k} \in L_{2} (p)}} \{ (L_{1}^{k} )^{{p_{1}^{k} }} \otimes (L_{2}^{k} )^{{p_{2}^{k} }} \}\).

  3. (3)

    \(\lambda L(p) = \cup_{{L^{k} \in L(p)}} \lambda p^{k} L^{k} ,\lambda \ge 0.\)

  4. (4)

    \(L(p)^{\lambda } = \cup_{{L^{k} \in L(p)}} {{\{ }}(L^{k} {)}^{{\lambda p^{k} }} {{\} }}{.}\)

Definition 2

[13] Let L(p) be a PLTS, \(L(p) = \{ L^{k} (p^{k} ){|}k = 1,2,...,\# L(p)\}\), then the score function of L(p) is given by.

$$ E(L(p)) = \sum\limits_{k = 1}^{\# L(p)} {L^{k} } p^{k} /\sum\limits_{k = 1}^{\# L(p)} {p^{k} } . $$
(2)

Definition 3

[13] Let L(p) be a PLTS, \(L(p) = \{ L^{k} (p^{k} ){|}k = 1,2,...,\# L(p)\}\), then the deviation degree of L(p) is given by.

$$ \sigma (L(p)) = \frac{{\sqrt {\sum\nolimits_{k = 1}^{\# L(p)} {(p^{k} L^{k} - \overline{\alpha })^{2} )} } }}{{\sum\nolimits_{k = 1}^{\# L(p)} {p^{k} } }}, $$
(3)

where \(\overline{\alpha } = \sum\nolimits_{k = 1}^{\# L(p)} {L^{k} p^{k} /\sum\nolimits_{k = 1}^{\# L(p)} {p^{k} } }\), \(\# L(p)\) is the number of terms in \(L(p)\).

The comparison rules of PLTSs based on score function and deviation degree are as follows:

(1) If \(E(L_{1} (p)) > E(L_{2} (p))\), then \(L_{1} (p) < L_{2} (p).\)

(2) If \(E(L_{1} (p)) < E(L_{2} (p))\), then \(L_{1} (p) > L_{2} (p).\)

(3) If \(E(L_{1} (p)) = E(L_{2} (p))\), then.

  1. 1)

    If \(\sigma (L_{1} (p)) > \sigma (L_{2} (p))\), then \(L_{1} (p) < L_{2} (p).\)

  2. 2)

    If \(\sigma (L_{1} (p)) < \sigma (L_{2} (p))\), then \(L_{1} (p) > L_{2} (p).\)

  3. 3)

    If \(\sigma (L_{1} (p)) = \sigma (L_{2} (p))\), then \(L_{1} (p) = L_{2} (p).\)

To aggregate multiple PLTSs, the most used approach is aggregation operators. In Pang’s research [13], some aggregation operators are introduced to integrate PLTSs and the definition of probabilistic linguistic weighted averaging (PLWA) operator is given as follows.

Definition 4

[13] Let \(L_{i} (p) = \{ L_{i}^{k} (p_{i}^{k} )|k = 1,2,...,\# L(p)\} (i = 1,2,...,n)\) be n PLTSs, then the PLWA operator is.

$$ \begin{aligned} &PLWA(L_{1} (p),L_{2} (p),...,L_{n} (p)) \\&\quad = w_{1} L_{1} (p) \oplus w_{2} L_{2} (p)\oplus ... \oplus w_{n} L_{n} (p) \hfill \\ &\quad = \cup_{{L_{{1}}^{{\text{k}}} \in L_{1} (p)}} \{ w_{1} p_{1}^{k} L_{1}^{k} \} \oplus \cup_{{L_{{2}}^{{\text{k}}} \in L_{2} (p)}} \{ w_{2} p_{2}^{k} L_{2}^{k} \} \\ & \quad\quad \oplus ... \oplus \cup_{{L_{{\text{n}}}^{{\text{k}}} \in L_{n} (p)}} \{ w_{n} p_{n}^{k} L_{n}^{k} \} , \hfill \\ \end{aligned} $$
(4)

where \(L_{i}^{k}\) and \(p_{i}^{k}\) denote the kth term and its probability in Li(p), and wi (i = 1,2,…,n) is the weight of Li(p).

Incomplete decision matrix

In decision making, above of all, DMs give their evaluations for distinct alternatives associated with multiple attributes according to their knowledge and experience. Given PLTS is a useful and superior tool to express experts’ opinions, especially in some emergency and vital decision cases, it becomes the preferred expression model. To describe experts’ evaluation information, the decision matrices are used as follows:

$$ P^{1} = \left[ {\begin{array}{*{20}c} {L_{11} } & {L_{12} } & {...} & {L_{1n} } \\ {L_{21} } & {L_{22} } & {...} & {L_{2n} } \\ {...} & {...} & {...} & {...} \\ {L_{m1} } & {L_{m2} } & {...} & {L_{mn} } \\ \end{array} } \right]\,P^{2} = \left[ {\begin{array}{*{20}c} {L_{11} } & * & {...} & {L_{1n} } \\ * & {L_{22} } & {...} & {L_{2n} } \\ {...} & {...} & {...} & {...} \\ {L_{m1} } & * & {...} & {L_{mn} } \\ \end{array} } \right]. $$

In decision matrix \(P^{1}\) represented by PLTS,\(L_{ij}\) represents the evaluation value of alternative i over attribute j,\(L_{ij} = \{ L_{ij}^{k} (p_{ij}^{k} )|k = 1,2,...,\# L_{ij} (p_{ij} )\}\). If all \(L_{ij}\) are known,\(P^{1}\) is called the complete decision matrix. However, due to some reasons, DMs cannot give a complete decision matrix in most cases. In decision matrix \(P^{2}\), ‘*’ denotes the missing value and \(P^{2}\) is called the incomplete decision matrix because there exist a values in \(P^{2}\). The element with missing value is not equal to useless or valueless information, which should not be ignored [14]. Thus, before conducting the decision making, the process to estimate missing information in decision matrix is an indispensable segment. For the sake of convenience, some concepts and mathematical notions are given as follows:

a. Missing expert: the expert whose decision matrix is incomplete. The set of missing experts is denoted as \(ME = \{ e_{k}^{*} |k = 1,2,...,\# e^{*} \}\), where \(\# e^{*}\) is the number of missing experts.

b. Referenced expert: the expert that they provide reference evaluation information for missing expert. The set of referenced experts is denoted as \(RE = \{ e_{k}^{r} |k = 1,2,...,\# e^{r} \}\), where \(\# e^{r}\) is the number of referenced experts.

c. Missing element: the element information in decision matrix is unknown. Denote \(L_{ijk}^{*} (p)\) as missing element, which means that the information of expert ek over alternative i associated with attribute j is missing. The set of missing elements is denoted as \(ML = \{ L_{ij}^{*} |i = 1,2,...,m;j = 1,2,...,n\}\).

Estimating incomplete PLTS information based on knowledge-match degree with reliability

When estimating incomplete decision matrix, an important approach is to refer to evaluation information from other experts (namely referenced experts). Owing to distinct knowledge levels, referenced experts provide the evaluation information of different quality in which the information quality includes two indicators, accuracy and reliability. The higher the accuracy and reliability of evaluation information, the bigger the importance of referenced expert. Thus, in this section, a new method to estimate missing PLTS information in complete decision matrix is proposed based on knowledge-match degree with reliability and an algorithm is designed to work out the approach.

Knowledge-match degree with reliability

To measure the accuracy of evaluation information of referenced expert, a new indicator is proposed, namely knowledge-match degree. The main idea is that the collective evaluation is viewed as benchmark, then the knowledge-match degree can be obtained according to the distance between individual evaluation and the benchmark. The bigger the distance, the low is the knowledge-match degree. The distance is given by

$$ d_{ij}^{k} = {|}E(L_{ijk} ) - \frac{1}{{\# e^{r} }}\sum\limits_{{g = {1}}}^{{{{\# }}e^{r} }} {E(L_{ijg} )} {|,} $$
(5)

where \(d_{ij}^{k}\) denotes the distance between the referenced expert \(e_{k}^{r}\) and collective evaluation over alternative i associated with attribute j. \(\# e^{r}\) is the number of referenced experts in RE.

Definition 5

Suppose that \(e_{k}^{r}\) is a referenced expert and RE is the set of referenced experts, \(e_{k}^{r} \in RE\),\(RE = \{ e_{k}^{r} |k = 1,2,...,\# e^{r} \}\). With respect to the evaluation information of \(e_{k}^{r}\) \(L_{ijk} (p)\), its knowledge-match degree is given by.

$$ KN_{ij}^{k} = 1 - \frac{{d_{ij}^{k} }}{{\sum\nolimits_{g = 1}^{{\# e^{r} }} {d_{ij}^{g} } }},g = 1,2,...,\# e^{r} , $$
(6)

where \(KN_{ij}^{k}\) denotes the knowledge-match degree of expert \(e_{k}^{r}\) over alternative i associated with attribute j. \(\# e^{r}\) is the number of referenced experts in RE.

Remarks

As we see, the knowledge-match degree is used to measure the accuracy of referenced experts’ evaluations. In the context of PLTS, experts provide their evaluation information by PLTS. When calculating the knowledge-match degree, the PLTS is transformed to a crisp value by score function (Eq. (2)). The benchmark is represented by the mean of collective evaluations, namely \(\frac{1}{{\# e^{r} }}\sum\nolimits_{{g = {1}}}^{{{{\# }}e^{r} }} {E(L_{ijg} )}\) in Eq. (5). In Eq. (6), we conduct a normalization of distance so as to obtain the knowledge-match degree which belongs to between 0 and 1.

Example 1

Let S = {s0, s1, s2, s3, s4} be the LTS and the set of referenced experts is \(RE = \{ e_{k}^{r} |k = 1,2,3\}\), the evaluations with respect to alternative 1 associated with attribute 1 are:

\(L_{11}^{1} (p) = \{ s_{0} (0.3),s_{1} (0.7)\}\), \(L_{11}^{2} (p) = \{ s_{1} (0.5),s_{2} (0.5)\}\), \(L_{11}^{3} (p) = \{ s_{1} (1)\} .\)

Then, by Eq. (5), we can get the distance:\(d_{11}^{1} = 0.43\), \(d_{11}^{2} = 0.49\), \(d_{11}^{3} = 0.08\).

According to Eq. (6), the knowledge-match degrees of \(e_{k}^{r}\) are.

$$KN_{11}^{1} = 0.57, KN_{11}^{2} = 0.51 KN_{11}^{3} = 0.92.$$

Besides, reliability is also an essential indicator to measure the importance of evaluations. In some cases, expert may give the evaluations with high knowledge-match degree but the expert has great uncertainty. Thus, on the basis of knowledge-match degree, the reliability of evaluation information should be also considered. Evidently, the higher the reliability, the bigger the importance of the evaluation information provided by the referenced expert. In the context of PLTS, there exists fuzziness in a PLTS and the more the fuzziness, the more the uncertainty of referenced expert. Hence, the reliability of evaluation information can be obtained by calculating fuzzy entropy of PLTS. Evidently, the smaller the fuzzy entropy, the bigger the reliability. In this research, we adopt the method proposed by Liu [29] to measure fuzzy entropy of PLTS and give the definition of reliability.

Definition 6

[29] Suppose that S is a LTS,\(S = \{ s_{\alpha } {|}\alpha = 0,1,2,...,2\tau \}\). Let \(L(p) = \{ L^{k} (p^{k} ){|}k = 1,2,...,\# L(p)\}\) be a PLTS, then the fuzzy entropy of L(p) is given by.

$$ E_{F} (L(p)) = 1 - \sum\limits_{i = 1}^{\# L(p)} {p_{i} } |1 - 2\alpha_{i} |, $$
(7)

where EF denotes the fuzzy entropy, \(\alpha_{i} = L_{i} /(2\tau )\), and g represents the number of terms in L(p).

Evidently, the value of fuzzy entropy (Eq. (7)) is between 0 and 1 (contain 0,1). With the aid of fuzzy entropy, we propose the definition of PLTS reliability.

Definition 7

Suppose that S is a LTS, and \(S = \{ s_{\alpha } {|}\alpha = 0,1,2,...,2\tau \}\). Let \(L(p) = \{ L^{k} (p^{k} ){|}k = 1,2,...,\# L(p)\}\) be a PLTS, then the reliability of L(p) is given by.

$$ R(L(p)) = 1 - E_{F} . $$
(8)

According to Eqs. (7), (8) is also written as

$$ R(L(p)) = \sum\limits_{i = 1}^{\# L(p)} {p_{i} |1 - 2\alpha_{i} |} . $$
(9)

Proposition 1.

In definition 7, the value of reliability is always between 0 and 1 (contain 0,1).

Proof.

The formula of reliability is Eq. (9), and \(R(L(p)) = \sum\limits_{i = 1}^{\# L(p)} {p_{i} |1 - 2\alpha_{i} |}\). For \(\alpha_{i} = L_{i} /(2\tau )\), Li is the subscript of the i th term in L(p). Evidently, \(0 \le L_{i} /(2\tau ) \le 1\), namely \(0 \le \alpha_{i} \le 1\). Thus, \(0 \le { |1 - 2}\alpha_{i} {| } \le {1}\). Since \(\sum\nolimits_{i = 1}^{\# L(p)} {p_{i} = 1}\), \(0 \le \sum\nolimits_{i = 1}^{\# L(p)} {p_{i} |1 - 2\alpha_{i} |} \le 1\). Hence, \(R(L(p)) \in [0,1]\).

Combining the knowledge-match degree with reliability, we can well measure the reliability level of evaluation information provided by referenced experts. The knowledge-match degree is used to identify the accuracy of evaluations while the reliability checks whether the evaluation information is reliable. The knowledge-match degree with reliability is given by

$$ KR_{ij}^{k} = KN_{ij}^{k} \cdot R_{ij}^{k} , $$
(10)

where \(KR_{ij}^{k}\) represents the knowledge-match degree with reliability provided by expert \(e_{k}^{r}\) with respect to alternative i associated to attribute j. Since \(KN_{ij}^{k} \in [0,1]\) and \(RE_{ij}^{k} \in [0,1]\), \(KR_{ij}^{k}\) also belongs to the interval [0,1].

Example 2

According to example 1, \(L_{11}^{1} (p) = \{ s_{0} (0.3),s_{1} (0.7)\}\), \(L_{11}^{2} (p) = \{ s_{1} (0.5),s_{2} (0.5)\}\), and \(L_{11}^{3} (p) = \{ s_{1} (1)\}\), and their knowledge-match degrees are \(KN_{11}^{1} = 0.57\), \(KN_{11}^{2} = 0.51\), and \(KN_{11}^{3} = 0.92\).

By Eq. (9), the reliability is \(R_{11}^{1} = 0.65\), \(R_{11}^{2} = 0.25\), and \(R_{11}^{3} = 0.5\).

According to Eq. (10), we can obtain the knowledge-match degree with reliability:

$$KR_{11}^{1} = 0.3705, KR_{11}^{2} = 0.1275, KR_{11}^{3} = 0.4600.$$

Estimating incomplete decision matrix

Different from Liang’s method [20] to estimate the missing values, we consider the evaluation information of referenced experts and utilize the knowledge-match degrees with reliability to aggregate the referenced information. The knowledge-match degree with reliability can be obtained by Eq. (10), namely \(KR_{ij}^{k} = KN_{ij}^{k} \cdot R_{ij}^{k}\), while it should be normalized first to become the aggregation operator. The normalized \(KR_{ij}^{k}\) is

$$ NKR_{ij}^{k} = \frac{{NKR_{ij}^{k} }}{{\sum\nolimits_{\begin{subarray}{l} g = 1 \\ g \ne k \end{subarray} }^{{\# e^{r} }} {NKR_{ij}^{g} } }}, $$
(11)

where \(\# e^{r}\) is the number of referenced experts in RE.

Subsequently, according to Eq. (4), namely the PLWA operator, obtain the estimated value of missing information \(L_{ij}^{*}\) by aggregating all referenced evaluations of referenced experts. Among that, \(NKR_{ij}^{k}\) is viewed as the weight of PLWA operator. A missing expert \(e_{k}^{*}\) may have several missing values in the decision matrix. In this case, estimate all missing values of \(e_{k}^{*}\) and then change to the next missing expert. Thus, we design an algorithm to estimate the missing information.

figure a

The application of estimation method in LSGDM

Problem description

In the LSGDM, the purpose is to achieve the optimal scheme in all alternatives. In our research, PLTS is used to express the evaluation information of DMs. If there exist incomplete PLTS information in decision matrix, the missing information will be first estimated. In the process of LSGDM, some notions are as follows:

The alternative set: \(X = \{ x_{i} |i = 1,2,...,m\}\).

The attribute set: \(C = \{ c_{j} |j = 1,2,...,n\}\).

The expert set: \(E = \{ e_{k} |k = 1,2,...,p\}\).

The set of missing experts: \(ME = \{ e_{k}^{*} |i = 1,2,...,\# e^{*} \}\).

The set of referenced experts: \(RE = \{ e_{k}^{r} |k = 1,2,...,\# e^{r} \}\).

The set of missing elements: \(ML = \{ L_{ij}^{*} |i = 1,2,...,m;j = 1,2,...,n\}\).

In this study, it contains four main steps to solve the problem of LSGDM including (1) estimating incomplete information, (2) subgroup detecting, (3) consensus reaching process (CRP), and (4) selection process. In step 1, to handle the incomplete PLTS information in decision matrices provided by experts, the Algorithm 1 is applied. In step 2, subgroup detecting means that the experts are divided into some small subgroups in term of specific standards or methods to simplify the complexity of decision problem. In step 3, CRP is an important section to check whether the consensus of experts satisfies the expected level (consensus threshold). In step 4, all evaluation information from subgroups are aggregated to obtain the comprehensive evaluation value. According to the comprehensive values, rank the alternatives.

The process of LSGDM based on the estimation of incomplete decision matrix

In this subsection, based on the proposed estimation method, the decision process is developed to solve the problem of LSGDM. To see the whole framework intuitively, we give the flowchart of LSGDM (Fig. 1).

Fig. 1
figure 1

Flowchart of LSGDM

Estimating incomplete information

In the beginning of decision, experts give their evaluations over various alternatives associated to attributes in term of professional knowledge and experience. However, there always exist the incomplete evaluation information provided by DMs because of the limited time or the complexity of decision problem. To efficiently utilize the evaluation information, it is vital to fill the missing elements in decision matrices. According to the Algorithm 1, the missing values can be estimated with the aid of the related information of referenced experts. The details of estimation are omitted here.

Subgroup detecting

In traditional group decision making, the number of DMs may be less and it is often in the interval [3, 20]. Nevertheless, the number of experts participating decision making is large in LSGDM. The decision information given by them are very complicated and it is much difficult to analysis if we directly conduct the decision making. Thus, all experts are divided to some small subgroups to simplify the decision analysis. K-means clustering is a useful method to classify and it has been successfully applied in LSGDM to solve subgroup detecting [3]. In this research, we also adopt this method to divide experts.

With respect to K-means method, two important indicators that should be determined are K value and initial centroids For the K value, we mainly obtain it according to the error sum of squares SSE. By means of software, such as Python, draw graphic of relation between SSE and K value, and find out the point in which the degree of decline becomes smooth from sharp, which is the best K value. Then, we randomly choose K experts as the initial centroids and remaining experts are assigned to the nearest centroids. The process of subgroup detecting is summarized in Algorithm 2 as follows.

figure b

Consensus reaching process based on trust relationship

Here, a consensus reaching process is developed to ensure that experts in each subgroups are able to reach the accepted consensus level. Since the complete consensus is difficult to achieve and it will consume large cost, the “soft consensus” is often used to measure whether the evaluations of experts satisfy the expected consensus. That is to say, we preset the consensus threshold \(\xi\), and it reaches consensus if the consensus index \(CI \ge \xi\). However, in most cases, it is difficult to reach consensus at the beginning as there always exist some conflicts of opinions given by DMs, which leads to low consensus degree. If \(CI < \xi\), some decision matrices with low consensus will be revised to improve the consensus level. The consensus reaching process (CRP) contains three parts: consensus measuring, identification strategy and consensus improvement.

Consensus measuring

When measuring the consensus of subgroup, Li et al. [30] considered it from four aspects: element, alternative, decision matrix and subgroup. In Li’s research, first, the collective opinion needs to be obtained by aggregating the information of all subgroups. However, it may be complicated in calculation by utilizing Li’s method. Thus, we directly measure the consensus degree in terms of distance between any two elements, which does not need to calculate the collective opinion in advance. After measuring the consensus level of element, alternative, decision matrix and subgroup respectively, the collective consensus degree can be obtained. The process of consensus measure is as follows:

  1. a.

    The consensus of element level

    $$ \begin{aligned}CI(L_{ij}^{{k_{1} }} )& = \frac{1}{{{{\# }}G_{t} - 1}}\sum\limits_{\begin{subarray}{l} k_{2} = 1 \\ k_{2} \ne k_{1} \end{subarray} }^{{\# G_{t} }}\\&\quad\times {(1 - |E(L_{ij}^{{k_{1} }} (p)) - E(L_{ij}^{{k_{2} }} (p))|} ), \, e_{{k_{1} }} \in G_{t} . \end{aligned}$$
    (14)
  2. b.

    The consensus of alternative level

    $$ CI(L_{i}^{k} ) = \frac{1}{n}\sum\limits_{j = 1}^{n} {CI(L_{ij}^{k} )} . $$
    (15)
  3. c.

    The consensus of decision matrix level

    $$ CI(P^{{k_{1} }} ) = \frac{1}{m}\sum\limits_{i = 1}^{m} {CI(L_{i}^{{k_{1} }} )} $$
    (16)
  4. d.

    The consensus of subgroup level

    $$ CI(G_{t} ) = \frac{1}{{\# G_{t} }}\sum\limits_{k = 1}^{{\# G_{t} }} {CI(P^{k} )} . $$
    (17)

Finally, the collective consensus is

$$ CI = \frac{1}{K}\sum\limits_{t = 1}^{K} {CI(G_{t} )} . $$
(18)

Identification strategy

If \(CI \ge \xi\), it satisfies the expected consensus and continue to perform to the next step, selection process. But if \(CI < \xi\), we should identify the elements with low consensus and revise them to reach the consensus threshold. It will consume large cost and largely change original information of experts to modify all elements which do not satisfy the consensus threshold. Thus, with the aid of Li’s research [30], we only need to revise the element with the lowest consensus. The identification processes are as follows:

  1. a.

    Identify the subgroup with the lowest consensus

    $$ AL(G_{t} ) = \{ G_{t} \left| {CI_{t} = \min CI(G_{t} )} \right.\} , $$
    (19)

    where \(AL(G_{t} )\) represents the subgroup with the lowest consensus among all subgroups.

  2. b.

    Identify the decision matrix with lowest consensus

    $$ AL(P^{k} ) = \{ P^{k} \left| {CI(P^{k} ) = \min CI(P^{k} )} \right. \wedge e_{k} \in AL(G_{t} )\} . $$
    (20)
  3. c.

    Identify the alternative with the lowest consensus

    $$ AL(L_{i}^{k} ) = \{ L_{i}^{k} \left| {CI(L_{i}^{k} ) = \min CI(L_{i}^{k} )} \right. \wedge L_{i}^{k} \in AL(P^{k} )\} . $$
    (21)
  4. d.

    Identify the element with the lowest consensus

    $$\begin{aligned} AL(L_{ij}^{k} ) = \{ L_{ij}^{k} | CI(L_{ij}^{k} ) = \min CI(L_{ij}^{k} ) \wedge L_{ij}^{k} \in AL(L_{ij}^{k})\}.\end{aligned}$$
    (22)

After conducting the above four steps, we can obtain the final element \(AL(L_{ij}^{k} )\) which needs to be modified. Evidently, through the identification strategy, it only needs to adjust the \(AL(L_{ij}^{k} )\) if the consensus level does not satisfy consensus threshold. The advantage of the method is to save the adjustment cost and remain the original evaluation information to the greatest extent.

Consensus improvement based on trust relationship

For the element \(AL(L_{ij}^{k} )\), it is the evaluation information given by expert \(e_{k}\) with respect to alternative i associated to attribute j. To some extent, the initial information will be changed when modifying the element with the lowest consensus. To reduce the loss of original evaluation information, we develop a new method to modify the element based on trust relationship of experts. As we all know, in a social network, the higher the trust degree, the bigger the preference similarity of two experts. Here, we utilize social network analysis (SNA) and graph theory to represent the relationship of experts. The related basic knowledge are as follows.

Definition 8

[31] Let G (V, E, W) be a graph, where V is the set of vertices, \(V = \{ v^{k} |k = 1,2,...,p\}\), E denotes the set of edges, \(E = \{ e^{kl} |k \ne l \wedge k,l = 1,2,...,p\}\) and W represents the set of weights, \(W = \{ w^{kl} |k \ne l \wedge k,l = 1,2,...,p\}\), \(w^{kl} \in [0,1]\). The social network graph is shown in Fig. 2.

Fig. 2
figure 2

Social network graph

In this research, \(v^{k}\) denotes expert \(e_{k}\) and \(e^{kl}\) denotes the relationship between \(e_{k}\) and \(e_{l}\). \(w^{kl}\) represents the trust degree between \(e_{k}\) and \(e_{l}\). Here, we utilize the directional social work [22], namely \(e^{kl} \ne e^{lk}\) and \(w^{kl} \ne w^{lk}\). When analysis the trust relationship among experts, we denote the weight \(w^{kl}\) as the trust degree \(T_{kl}\) from \(e_{k}\) to \(e_{l}\).

Based on SNA, we can find out the most trusted expert of \(e_{k}\) in the same subgroup, which is denoted as \(e_{k}^{T}\),

$$ e_{k}^{T} = \{ e_{l} |T_{kl} = \max T_{kl} \wedge CI(P^{l} ) \ge \xi \wedge e_{k} ,e_{l} \in G_{t} \} . $$
(23)

In Eq. (23), the expert \(e_{k}^{T}\) needs satisfy three conditions. First, \(e_{k}\) and \(e_{k}^{T}\) are in the same subgroup. Second, the consensus degree of \(e_{k}^{T}\) must exceed the consensus threshold. Finally, the trust degree between \(e_{k}\) and \(e_{k}^{T}\) is the maximum one. According to the corresponding element of \(e_{k}^{T}\), we modify the \(AL(L_{ij}^{k} )\) and preset a parameter \(\theta\) to control the adjustment proportion. The modification mechanism is given by

$$ AL(\overline{L}_{ij}^{k} ) = (1 - \theta )L_{ij}^{k} (p) + \theta L_{ij}^{l} (p), $$
(24)

where \(AL(\overline{L}_{ij}^{k} )\) represents the modified \(AL(L_{ij}^{k} )\), \(L_{ij}^{l} (p) = \{ L_{ij}^{l} |e_{l} = e_{k}^{T} \}\).

By Eqs. (23) and (24), we can improve the consensus levels of subgroup and collective through modifying the elements with the low consensus iteratively. To realize the complete consensus reaching process, the Algorithm 3 is designed as follows.

figure c

Selection process

After the consensus reaching process, the evaluation information of all experts satisfy the consensus threshold. Subsequently, in the selection process, the decision matrices of experts will be aggregated into the collective decision matrix. Step by step, we first integrate the decision information in the subgroup by Eq. (4), where the weights of experts in the same subgroup are identical by default, e.g., the expert’s weight 1/#Gt. We denote the decision matrix of subgroup as \(P^{{G_{t} }}\):

$$ P^{{G_{t} }} = PLWA(P^{1} ,P^{2} ,...,P^{{\# G_{t} }} ). $$
(25)

Then, according to each decision matrix of subgroup \(P^{{G_{t} }}\), we can get the collective evaluation. The weights of subgroups \(\lambda_{t}\) are given by the size of persons in subgroup, e.g., \(\lambda_{t} = \# G_{t} /\sum\limits_{r = 1}^{K} {\# G_{r} }\). Thus, the collective decision matrix \(P^{c}\) is

$$ P^{c} = PLWA(P^{{G_{1} }} ,P^{{G_{2} }} ,...,P^{K} ). $$
(26)

As for the attribute weight \(\omega_{j}\), we preset the value in terms of the actual decision situation. Similarly, utilize PLWA and score function (Eq. (2)) to aggregate the attribute information of \(P^{c}\) into the comprehensive values of alternatives \(E(L_{i} )\). According to the comprehensive values, we rank the alternative in descending order and obtain the optimal one.

In our research, the main steps of LSGDM are summarized as follows:

Step 1. Collect the experts’ evaluation information and estimate the missing PLTS information in incomplete decision matrices. Using the Algorithm 1, the incomplete decision matrices can be transformed to the complete ones.

Step 2. Subgroup detecting. According to the Algorithm 2, the experts are divided into some distinct subgroups.

Step 3. Consensus reaching process. By utilizing the Algorithm 3, the evaluations of experts are able to satisfy anticipated consensus level.

Step 4. Selection process. According to Eqs. (25) and (26), obtain the collective decision matrix \(P^{c}\). Calculate the comprehensive values of alternatives \(E(L_{i} )\) and rank the alternatives. Finally, we obtain the optimal one.

In next section, the proposed process of LSGDM will be applied to a medical case.

Case study

In medical field, decision analysis is often used to solve some complex decision problem, such as selection of drug suppliers, drug procurement management, medicine classification based on AI [32] and so on. When coping with such decision problems, large number of human experts who come from different fields and have rich experience are invited to give evaluations over distinct alternatives. Although human are multi-modal thinking and can make decision intelligently and efficiently according to their knowledge and experience [33], there often exist cognitive fuzziness and hesitation when people face the complex problem and this uncertainty should be considered. Thus, PLTS is a very useful tool to express the evaluation information of experts. Now, suppose that a hospital is facing such a decision problem:

There are four kinds of drugs: medicine A, medicine B, medicine C and medicine D. Since the four drugs can treat the same disease while they have different efficacy and side effects, the hospital will choose the best one among the four drugs. Twenty experts from various medical fields are invited to participate in this decision making. They will give evaluations by means of PLTS according to three evaluation criteria: efficacy, side effects and cost. Given the importance of different criteria, we preset the weights of attributes to 0.5, 0.3 and 0.2, respectively. According to the proposed decision process in “The application of estimation method in LSGDM”, we solve the problem step by step.

Evidently, this is a LSGDM. Experts use a linguistic term set with 5 granularity to give evaluations, namely S = {s0, s1, s2, s3, s4}. From left to right, the terms mean: very bad, bad, medium, good, and very good. There are totally 20 DMs, 4 alternatives (4 kinds of drugs) and 3 attributes (3 evaluation criteria) which are denoted as.

The set of alternatives: \(X = \{ x_{i} |i = 1,2,3,4\}\).

The attribute set: \(C = \{ c_{j} |j = 1,2,3\}\).

The weight vector of attribute: \(\omega = (0.5,0.3,0.2)\).

The expert set: \(E = \{ e_{k} |k = 1,2,...,20\}\).

Step 1. By means of PLTS, the experts give their evaluation information in the form of matrix as follows: (‘*’ represents missing information)

$$ P^{1} = \left[ {\begin{array}{*{20}c} {\{ s_{2} (1)\} } & {\{ s_{1} (0.3),s_{2} (0.7)\} } & {\{ s_{3} (1)\} } \\ {\{ s_{4} (0.8)\} } & {\{ s_{2} (1)\} } & {\{ s_{2} (0.6),s_{3} (0.2)\} } \\ {\{ s_{0} (0.8),s_{1} (0.2)\} } & {\{ s_{3} (0.9)\} } & {\{ s_{1} (0.5),s_{2} (0.5)\} } \\ {\{ s_{3} (1)\} } & {\{ s_{2} (1)\} } & {\{ s_{3} (0.8)\} } \\ \end{array} } \right] $$
$$ P^{2} = \left[ {\begin{array}{*{20}c} * & {\{ s_{2} (1)\} } & {\{ s_{1} (0.2),s_{2} (0.8)\} } \\ {\{ s_{2} (0.5),s_{3} (0.5)\} } & {\{ s_{2} (0.8)\} } & {\{ s_{3} (1)\} } \\ {\{ s_{2} (0.8)\} } & {\{ s_{4} (1)\} } & {\{ s_{1} (0.8),s_{2} (0.2)\} } \\ {\{ s_{3} (1)\} } & {\{ s_{2} (0.8),s_{3} (0.2)\} } & {\{ s_{2} (1)\} } \\ \end{array} } \right] $$
$$ \begin{aligned} &P^{3} = \left[ {\begin{array}{*{20}c} {\{ s_{3} (0.8)\} } & {\{ s_{2} (1)\} } & {\{ s_{2} (0.5),s_{3} (0.5)\} } \\ {\{ s_{3} (0.9),s_{4} (0.1)\} } & {\{ s_{1} (1)\} } & {\{ s_{3} (1)\} } \\ {\{ s_{4} (1)\} } & {\{ s_{1} (0.3),s_{2} (0.5)\} } & {\{ s_{2} (1)\} } \\ {\{ s_{2} (1)\} } & {\{ s_{3} (0.8)\} } & {\{ s_{1} (0.5),s_{2} (0.5)\} } \\ \end{array} } \right] \\ &P^{4} = \left[ {\begin{array}{*{20}c} {\{ s_{2} (0.5),s_{3} (0.5)\} } & * & {\{ s_{3} (1)\} } \\ {\{ s_{2} (0.8)\} } & {\{ s_{2} (0.3),s_{3} (0.7)\} } & {\{ s_{2} (1)\} } \\ {\{ s_{1} (0.4),s_{2} (0.6)\} } & {\{ s_{3} (1)\} } & {\{ s_{3} (0.2),s_{4} (0.8)\} } \\ {\{ s_{2} (1)\} } & {\{ s_{3} (1)\} } & {\{ s_{2} (0.5),s_{3} (0.5)\} } \\ \end{array} } \right] \end{aligned}$$
$$ P^{5} = \left[ {\begin{array}{*{20}c} {\{ s_{3} (1)\} } & {\{ s_{2} (0.5),s_{3} (0.5)\} } & {\{ s_{2} (1)\} } \\ {\{ s_{1} (0.5),s_{2} (0.3)\} } & {\{ s_{2} (1)\} } & {\{ s_{3} (0.9)\} } \\ {\{ s_{4} (1)\} } & {\{ s_{1} (0.6)\} } & {\{ s_{2} (1)\} } \\ {\{ s_{2} (1)\} } & {\{ s_{3} (1)\} } & {\{ s_{2} (0.6),s_{3} (0.4)\} } \\ \end{array} } \right] $$
$$ P^{6} = \left[ {\begin{array}{*{20}c} {\{ s_{1} (1)\} } & {\{ s_{2} (0.5),s_{3} (0.5)\} } & {\{ s_{3} (1)\} } \\ {\{ s_{3} (0.6),s_{4} (0.4)\} } & * & {\{ s_{2} (1)\} } \\ {\{ s_{2} (0.8)\} } & {\{ s_{3} (1)\} } & {\{ s_{2} (0.5),s_{3} (0.5)\} } \\ {\{ s_{1} (0.5)\} } & {\{ s_{2} (0.8),s_{3} (0.2)\} } & {\{ s_{2} (1)\} } \\ \end{array} } \right] $$
$$ P^{7} = \left[ {\begin{array}{*{20}c} {\{ s_{1} (0.8),s_{2} (0.2)\} } & {\{ s_{2} (1)\} } & {\{ s_{3} (1)\} } \\ {\{ s_{2} (0.6),s_{3} (0.4)\} } & {\{ s_{3} (1)\} } & {\{ s_{2} (0.5),s_{3} (0.5)\} } \\ {\{ s_{1} (0.8)\} } & {\{ s_{2} (0.6),s_{3} (0.4)\} } & {\{ s_{2} (1)\} } \\ {\{ s_{2} (1)\} } & {\{ s_{2} (1)\} } & {\{ s_{3} (1)\} } \\ \end{array} } \right] $$
$$ P^{8} = \left[ {\begin{array}{*{20}c} {\{ s_{2} (1)\} } & {\{ s_{1} (0.5),s_{2} (0.5)\} } & {\{ s_{3} (1)\} } \\ {\{ s_{2} (0.8),s_{3} (0.2)\} } & {\{ s_{1} (1)\} } & {\{ s_{1} (0.3),s_{2} (0.7)\} } \\ {\{ s_{3} (1)\} } & {\{ s_{2} (0.6),s_{3} (0.4)\} } & {\{ s_{2} (1)\} } \\ {\{ s_{2} (1)\} } & {\{ s_{2} (0.5),s_{3} (0.5)\} } & {*} \\ \end{array} } \right] $$
$$ P^{9} = \left[ {\begin{array}{*{20}c} {\{ s_{2} (0.8)\} } & {\{ s_{3} (1)\} } & {\{ s_{3} (0.8),s_{4} (0.2)\} } \\ {\{ s_{1} (1)\} } & {\{ s_{3} (1)\} } & {\{ s_{0} (0.9)\} } \\ {\{ s_{2} (1)\} } & {\{ s_{3} (0.6),s_{4} (0.4)\} } & {\{ s_{3} (0.8)\} } \\ {\{ s_{1} (0.3),s_{2} (0.7)\} } & {\{ s_{2} (0.8),s_{3} (0.2)\} } & {\{ s_{3} (1)\} } \\ \end{array} } \right] $$
$$ P^{10} = \left[ {\begin{array}{*{20}c} {\{ s_{3} (1)\} } & {\{ s_{2} (0.3),s_{3} (0.7)\} } & {\{ s_{4} (1)\} } \\ {\{ s_{1} (0.5),s_{2} (0.5)\} } & {\{ s_{2} (1)\} } & {\{ s_{2} (0.8),s_{3} (0.2)\} } \\ {\{ s_{1} (0.3),s_{2} (0.7)\} } & {\{ s_{3} (1)\} } & {\{ s_{4} (1)\} } \\ {\{ s_{2} (1)\} } & {\{ s_{2} (0.3),s_{3} (0.7)\} } & {\{ s_{2} (0.8)\} } \\ \end{array} } \right] $$
$$ P^{11} = \left[ {\begin{array}{*{20}c} {\{ s_{2} (1)\} } & {\{ s_{2} (0.5),s_{3} (0.5)\} } & {\{ s_{1} (1)\} } \\ * & {\{ s_{4} (0.8)\} } & {\{ s_{1} (0.3),s_{2} (0.7)\} } \\ {\{ s_{3} (0.3),s_{4} (0.7)\} } & {\{ s_{2} (1)\} } & {\{ s_{3} (1)\} } \\ {\{ s_{2} (1)\} } & {\{ s_{3} (0.6),s_{4} (0.4)\} } & {\{ s_{2} (1)\} } \\ \end{array} } \right] $$
$$ P^{12} = \left[ {\begin{array}{*{20}c} {\{ s_{1} (1)\} } & {\{ s_{2} (0.6),s_{3} (0.4)\} } & {\{ s_{4} (1)\} } \\ {\{ s_{1} (0.6),s_{2} (0.4)\} } & {\{ s_{2} (1)\} } & {\{ s_{2} (0.5),s_{3} (0.5)\} } \\ {\{ s_{2} (1)\} } & {\{ s_{3} (1)\} } & {\{ s_{1} (0.5),s_{2} (0.5)\} } \\ {\{ s_{3} (0.6),s_{4} (0.4)\} } & {\{ s_{2} (0.8),s_{3} (0.2)\} } & {\{ s_{2} (1)\} } \\ \end{array} } \right] $$
$$ P^{13} = \left[ {\begin{array}{*{20}c} {\{ s_{2} (1)\} } & {\{ s_{1} (0.8),s_{2} (0.2)\} } & {\{ s_{3} (1)\} } \\ {\{ s_{2} (0.4),s_{3} (0.6)\} } & {\{ s_{3} (1)\} } & {\{ s_{2} (0.8)\} } \\ {\{ s_{1} (1)\} } & {\{ s_{3} (1)\} } & {\{ s_{2} (0.6),s_{3} (0.4)\} } \\ {\{ s_{2} (0.5),s_{3} (0.5)\} } & {\{ s_{1} (1)\} } & {\{ s_{2} (1)\} } \\ \end{array} } \right] $$
$$ P^{14} = \left[ {\begin{array}{*{20}c} {\{ s_{1} (0.6),s_{2} (0.4)\} } & {\{ s_{2} (1)\} } & {\{ s_{3} (0.5),s_{4} (0.5)\} } \\ {\{ s_{3} (1)\} } & {\{ s_{1} (0.3),s_{2} (0.7)\} } & {\{ s_{2} (1)\} } \\ {\{ s_{2} (1)\} } & {\{ s_{3} (0.8)\} } & {\{ s_{2} (0.4),s_{3} (0.6)\} } \\ {\{ s_{4} (1)\} } & {\{ s_{2} (0.8),s_{3} (0.2)\} } & {\{ s_{3} (1)\} } \\ \end{array} } \right] $$
$$ P^{15} = \left[ {\begin{array}{*{20}c} {{{\{ }}s_{1} {\text{(1)\} }}} & {\{ s_{2} (0.5),s_{3} (0.5)\} } & {\{ s_{2} (1)\} } \\ {\{ s_{1} (0.6),s_{2} (0.4)\} } & {\{ s_{3} (0.5),s_{4} (0.3)\} } & {\{ s_{4} (1)\} } \\ {\{ s_{2} (1)\} } & {\{ s_{3} (1)\} } & {\{ s_{2} (0.5),s_{3} (0.5)\} } \\ {\{ s_{1} (0.6),s_{2} (0.4)\} } & {\{ s_{3} (0.8),s_{4} (0.2)\} } & {\{ s_{2} (1)\} } \\ \end{array} } \right] $$
$$ P^{16} = \left[ {\begin{array}{*{20}c} {\{ s_{2} (1)\} } & {\{ s_{3} (0.5),s_{4} (0.5)\} } & {\{ s_{3} (1)\} } \\ {\{ s_{3} (1)\} } & {\{ s_{2} (1)\} } & {\{ s_{2} (0.8),s_{3} (0.2)\} } \\ {\{ s_{1} (0.4),s_{2} (0.6)\} } & {\{ s_{3} (1)\} } & {\{ s_{2} (1)\} } \\ {\{ s_{1} (0.8)\} } & {\{ s_{2} (0.3),s_{3} (0.7)\} } & {\{ s_{4} (1)\} } \\ \end{array} } \right] $$
$$ P^{17} = \left[ {\begin{array}{*{20}c} {\{ s_{1} (0.5),s_{2} (0.5)\} } & {\{ s_{2} (1)\} } & {\{ s_{3} (1)\} } \\ {\{ s_{2} (0.8)\} } & {\{ s_{1} (0.6),s_{2} (0.4)\} } & {\{ s_{2} (1)\} } \\ {\{ s_{3} (1)\} } & {\{ s_{1} (0.5),s_{2} (0.5)\} } & {\{ s_{2} (0.6),s_{3} (0.4)\} } \\ {\{ s_{2} (1)\} } & {\{ s_{3} (1)\} } & {\{ s_{1} (0.5),s_{2} (0.5)\} } \\ \end{array} } \right] $$
$$ P^{18} = \left[ {\begin{array}{*{20}c} {\{ s_{2} (1)\} } & {\{ s_{1} (0.6),s_{2} (0.4)\} } & {\{ s_{3} (1)\} } \\ {\{ s_{2} (0.3),s_{3} (0.7)\} } & {\{ s_{2} (0.8)\} } & {\{ s_{2} (1)\} } \\ {\{ s_{1} (0.5),s_{2} (0.5)\} } & {\{ s_{3} (1)\} } & {\{ s_{2} (0.3),s_{3} (0.7)\} } \\ {\{ s_{2} (1)\} } & {\{ s_{2} (0.5),s_{3} (0.5)\} } & {\{ s_{2} (1)\} } \\ \end{array} } \right] $$
$$ P^{19} = \left[ {\begin{array}{*{20}c} {\{ s_{3} (1)\} } & {\{ s_{2} (0.9),s_{3} (0.1)\} } & {\{ s_{2} (1)\} } \\ {\{ s_{1} (0.3),s_{2} (0.7)\} } & {\{ s_{2} (1)\} } & {\{ s_{2} (0.5),s_{3} (0.5)\} } \\ {\{ s_{3} (0.8)\} } & {\{ s_{1} (0.4),s_{2} (0.6)\} } & {\{ s_{2} (1)\} } \\ {\{ s_{2} (1)\} } & {\{ s_{2} (0.9)\} } & {\{ s_{3} (1)\} } \\ \end{array} } \right] $$
$$ P^{20} = \left[ {\begin{array}{*{20}c} {\{ s_{1} (0.5),s_{2} (0.5)\} } & {\{ s_{3} (1)\} } & {\{ s_{2} (0.5),s_{3} (0.5)\} } \\ {\{ s_{2} (1)\} } & {\{ s_{1} (0.8)\} } & {\{ s_{1} (0.6),s_{2} (0.4)\} } \\ {\{ s_{3} (1)\} } & {\{ s_{1} (0.7),s_{2} (0.3)\} } & {\{ s_{3} (1)\} } \\ {\{ s_{2} (0.5),s_{3} (0.5)\} } & {\{ s_{2} (1)\} } & {\{ s_{1} (1)\} } \\ \end{array} } \right] $$

In the above decision matrices, find out the incomplete decision matrices and estimate the missing information. The missing experts and missing elements are as follows:

The set of missing experts: \(ME = \{ e_{2}^{*} ,e_{4}^{*} ,e_{6}^{*} ,e_{8}^{*} ,e_{11}^{*} \}\).

The set of missing elements: \(ML = \{ L_{11}^{*} ,L_{12}^{*} ,L_{22}^{*} ,L_{43}^{*} ,L_{21}^{*} \}\).

We take \(e_{2}^{*}\) as an example. For \(e_{2}^{*}\), the missing element is \(L_{11}^{*}\), namely denoting \(L_{112}^{*}\). According to the Algorithm 1, \(L_{112}^{*}\) can be estimated with the aid of PLTS information of referenced experts. In addition, the set of referenced experts is \(MR = \{ e_{1}^{r} ,e_{3}^{r} ,...,e_{20}^{r} \}\). Subsequently, calculate the knowledge-match degree with reliability on the element \(L_{11}\) for each referenced expert. By Eqs. (5)–(11), we can obtain the normalized \(KR_{ij}^{k}\):

\(NKR_{11}^{1} = 0\) \(NKR_{11}^{3} = 0.099\) \(NKR_{11}^{4} = 0.053\) \(NKR_{11}^{5} = 0.099\) \(NKR_{11}^{6} = 0.101\)

\(NKR_{11}^{7} = 0.081\) \(NKR_{11}^{8} = 0\) \(NKR_{11}^{9} = 0\) \(NKR_{11}^{10} = 0.099\) \(NKR_{11}^{11} = 0\)

\(NKR_{11}^{12} = 0.101\) \(NKR_{11}^{13} = 0\) \(NKR_{11}^{14} = 0.064\) \(NKR_{11}^{15} = 0.101\) \(NKR_{11}^{16} = 0\)

\(NKR_{11}^{17} = 0.053\) \(NKR_{11}^{18} = 0\) \(NKR_{11}^{19} = 0.099\) \(NKR_{11}^{20} = 0.053\)

Then, according to Eq. (4) and the \(NKR_{ij}^{k}\) is used as the aggregation weight. We can get the estimated value of missing element \(L_{112}^{*} = \{ s_{2} (0.4),s_{2.1} (0.6)\}\).

Similarly, for the rest elements in ML, we have

$$ L_{124}^{*} = \{ s_{2.3} (0.5),s_{2.8} (0.5)\} \,L_{226}^{*} = \{ s_{3} (1)\} $$
$$ L_{438}^{*} = \{ s_{1} (0.8),s_{1.6} (0.2)\} \,L_{2111}^{*} = \{ s_{2.8} (0.5),s_{3} (0.5)\} $$

Step 2. Subgroup detecting. Utilize the software Python to make a diagram between SSE and k value (see Fig. 3), and determine the best k value.

Fig. 3
figure 3

The diagram between SSE and k value

From Fig. 3, we can see that the best k value is equal to 3. Thus, the number of subgroup should be taken 3. According to the Algorithm 2, choose randomly the experts \(e_{1}\), \(e_{2}\) and \(e_{3}\) as initial centroids of subgroups. The result of subgroup detecting is.

Subgroup 1: \(e_{1}\),\(e_{4}\),\(e_{5}\),\(e_{9}\),\(e_{12}\).

Subgroup 2: \(e_{2}\),\(e_{6}\),\(e_{7}\),\(e_{11}\),\(e_{16}\),\(e_{17}\),\(e_{18}\).

Subgroup 3: \(e_{3}\),\(e_{8}\),\(e_{10}\),\(e_{13}\),\(e_{14}\),\(e_{15}\),\(e_{19}\),\(e_{20}\).

After three iterations, the subgroups are stable and the final result is as follows:

Subgroup 1: \(e_{1}\),\(e_{5}\),\(e_{9}\), \(e_{11}\),\(e_{12}\),\(e_{13}\).

Subgroup 2: \(e_{2}\),\(e_{4}\),\(e_{6}\),\(e_{7}\),\(e_{14}\),\(e_{16}\),\(e_{17}\),\(e_{18}\).

Subgroup 3: \(e_{3}\),\(e_{8}\),\(e_{10}\),,\(e_{15}\),\(e_{19}\),\(e_{20}\).

Step 3. Consensus reaching process according to the Algorithm 3. Given the importance of drug selection and the complexity of experts, we preset the consensus threshold \(\xi = 0.9\). By Eqs. (14) ~ (18), compute the consensus degrees from four aspects: element level, alternative level, decision matrix level and subgroup level. The consensus degrees on decision matrix level and subgroup level are shown in Table 2.

Table 2 Consensus degrees on different levels

According to the results in Table 1, we obtain the collective consensus degree CI = 0.8991. Since \(CI < \xi\), the collective consensus needs to be improved by modifying some elements. In term of the identification strategy, using Eqs. (19)–(22), we find out the element with lowest consensus. First, the subgroup with lowest consensus is G1 and the decision matrix level with lowest consensus is P12. Then, in P12, the alternative with lowest consensus is L1. Finally, the element with lowest consensus is L13 in the decision matrix of expert e12. Subsequently, based on the trust relationship, we revise the element L13. The social relationship among experts in G1 is shown in Fig. 4.

Fig. 4
figure 4

Social relationship of experts in G1

In Fig. 4, we see the most trusted expert of e12 is e9, namely \(e_{12}^{T} = \{ e_{9} \}\). In Eq. (24), \(\theta\) is set 0.2. Thus, \(AL(\overline{L}_{13}^{12} ) = (1 - 0.2)L_{13}^{12} (p) + 0.2L_{13}^{9} (p) = \{ s_{3.8} (0.8),s_{4} (0.2)\}\). Again, according to Eqs. (14) ~ (18), calculate the collective consensus degree CI = 0.9026. \(CI > \xi\), the consensus reaching process ends.

Step 4. Selection process. By Eq. (25), each decision matrix of subgroup can be obtained. According to the size of subgroup, the weights of subgroups are \(\lambda_{1} = 0.3\), \(\lambda_{2} = 0.4\) and \(\lambda_{3} = 0.3\). By Eq. (26), the collective decision matrix is

$$ P^{c} = \left[ {\begin{array}{*{20}c} {\{ s_{2.1} (0.6),s_{2.4} (0.4)\} } & {\{ s_{2.3} (0.5),s_{2.6} (0.5)\} } & {\{ s_{2.7} (0.3),s_{3} (0.7)\} } \\ {\{ s_{2} (0.2),s_{2.3} (0.8)\} } & {\{ s_{1.3} (0.8),s_{1.6} (0.2)\} } & {\{ s_{2} (0.7),s_{2.4} (0.3)\} } \\ {\{ s_{1.5} (0.6),s_{1.9} (0.4)\} } & {\{ s_{2} (0.7),s_{2.3} (0.3)\} } & {\{ s_{2} (0.8),s_{2.3} (0.2)\} } \\ {\{ s_{3.4} (0.5),s_{3.7} (0.5)\} } & {\{ s_{2} (0.5),s_{2.4} (0.5)\} } & {\{ s_{3} (0.4),s_{3.3} (0.6)\} } \\ \end{array} } \right]. $$

The weights of attributes are \(\omega_{1} = 0.5\),\(\omega_{2} = 0.3\) and \(\omega_{3} = 0.2\). By Eq. (4), the comprehensive evaluation of each alternative is

$$ L_{1} = \{ s_{2.3} (0.3),s_{2.4} (0.36),s_{2.5} (0.2),s_{2.6} (0.14)\} $$
$$ L_{2} = \{ s_{1.8} (0.112),s_{1.9} (0.524),s_{2} (0.316),s_{2.1} (0.048)\} $$
$$ L_{3} = \{ s_{1.8} (0.564),s_{1.9} (0.036),s_{2} (0.376),s_{2.1} (0.024)\} $$
$$ L_{4} = \{ s_{2.9} (0.1),s_{3} (0.25),s_{3.1} (0.4),s_{3.2} (0.25)\} . $$

According to Eq. (2), calculate the score functions of alternatives:

$$E(L_{1} ) \!=\! 2.418, E(L_{2} ) \!=\! 1.930, E(L_{3} ) \!=\! 1.886, E(L_{4} ) = 3.080$$

.

Thus, the priority of alternatives is: \(x_{4} \succ x_{1} \succ x_{2} \succ x_{3}\). Namely, the medicine D is the optimal drug.

Comparative analysis

In “Case study”, we use a medical case to illustrate our research method. Besides, the proposed method possesses more superiority and rationality from three aspects.

First and foremost, the method to estimate missing values in decision matrix is more scientific and cautious. Different from previous researches, we consider the accuracy and reliability of referenced information when they are used to estimated missing values, i.e., the knowledge-match degree with reliability, which can guarantee the accuracy of estimated values. However, in some previous methods, it has been never taken into account. For instance, Liang et al. [20] utilized the collaborative filtering algorithm to estimate missing values while it only considers the similarity of nearest neighbors. In some cases, even though the similarity of referenced information is the highest, it may be low accuracy if we do not take the accuracy and reliability of referenced information into account. Thus, it is necessary to use knowledge-match degree with reliability to measure the accuracy and dependability of referenced information.

Besides, in the process of consensus measuring, we directly calculate the distance between any two objectives to obtain the consensus degree \(CI(L_{ij}^{{k_{1} }} )\)(see Eqs. (14) ~ (18)), which can reduce the cost of time and calculation. However, in Li’s research, first, the collective opinion needs to be obtained by aggregating the information of all subgroups. It needs to conduct complicated calculations including the integration of decision matrix of each subgroup and collective decision matrix. Especially, in LSGDM, there exist large number of evaluation information. If we adopt Li’s method [30] to measure consensus level, the computed cost will be high. Hence, in this paper, the proposed method to measure consensus level is more simple and efficiency.

Finally, to improve the collective consensus, we design a new method to revise the elements based on trust relationship in the social network. Different from previous method, when modifying the element provided by the expert ek, we refer to the evaluation information of the most trusted expert of ek rather than all other trusted experts. That is to say, the proposed method can reduce the computation cost. Besides, the adjustment parameter \(\theta\) is proposed to control the amplitude of adjustment. The smaller is the parameter \(\theta\), the more the reserved original information of ek. According to actual situation and the trust degree, the parameter \(\theta\) can be set flexibly.

Conclusion

In large-scale group decision making, due to various reasons, DMs may provide incomplete decision matrices. The missing information in those incomplete decision matrices also play an important role. In addition, the PLTS is a useful tool to describe the information uncertainty. Thus, under the environment of LSGDM, we design a new method to estimate the missing PLTS information in incomplete decision matrices based on knowledge-match degree with reliability. Compared with previous methods, the proposed estimation method can enhance the accuracy and reliability of estimated values. Besides, to apply the estimation method to LSGDM, a complete process of LSGDM is developed and we propose a new consensus reaching process based on trust degree. The new CRP in LSGDM can well solve the consensus problem. In this paper, the case of medicine selection is used to illustrate our decision method, which indicates that the proposed decision process can help managers and decision makers to solve some complicated multi-attribute problem, especially the decision case with much uncertain information. In addition to its application in the field of medicine, the proposed method can well be applied to bid evaluation [34], logistics provider selection [35], emergency decision and so on.

However, there are some limitations in this research. The first defect is that the computation of estimation method may be large and the reliability of PLTS is obtained by means of simple fuzzy entropy, which need to be improved in next work. Besides, with respect to the process of consensus improvement, the adjustment parameter is set randomly, which should be determined scientifically and reasonably. Therefore, in the next work, we will solve the above problems. We will have a deep research on incomplete decision matrix in an uncertain environment and explore the application related to the PLTS decision methods.