1 Introduction

Yao (2009, 2010, 2011) introduced the notion of TWD in 2010s, which is focused on grouping all objects into three distinguishable and relevant sections and then design an appropriate strategy for each section, i.e. non-commitment, rejection, and acceptance. Different from the usual decision-making mode, TWD adds an uncertain part to better deal with the situation that it is hard to make an exact decision of acceptance or rejection directly with insufficient or incomplete information. As it is more in consistent with the characteristics of human mind, three-way decision has been attracted research interest of many researchers.

Up to now, there have been a lot of theoretical and applied achievements on TWD. Some scholars extended the traditional TWD on the basis of different situations. For example, Qian et al. (Qian et al. 2022) discussed sequential TWD problem in multi-granularity environment to obtain reliable decisions when there are several different granularity levels. Li et al. (2017a) established a TWD model with generalizations on the basis of subset evaluation. Yao (2018) proposed a model for three-way granular computing. A generalized multi-granularity sequential TWD model with multiple thresholds was developed by Qian et al. (2019). In order to realize the sorting and classification of intuitionistic fuzzy objects, Dai et al. (2023) constructed a TWD model with intuitionistic fuzzy concept consideration of the decision-maker’s preference. There has also been significant concern about the TWD procedure improvement. Chen et al. (2012) gave a method to find the threshold of TWD without setting the initial value. According to decision-theoretic rough sets with interval values, Liang and Liu (2014) put forward an optimum method for TWD to go a step further reduce the entire uncertainty. A method of aggregation for combining various TWD spaces was put forth by Hu et al. (2016). A sequential strategy was put up by Xu et al. (2020a) to address the issue of decision conflict by gradually introducing more precise knowledge. Based on criteria importance, Zhu et al. (2021) constructed a novel risk assessment framework to alter traditional failure mode and effect analysis method and aid maintenance supervision. The distinctive preponderance of TWD makes it popular in medical diagnosis (Chu et al. 2020; Yao and Azam 2015), decision support (Liang and Liu 2015; Zhang et al. 2020a), recommender system design (Zhang and Min 2016), classification (Xu et al. 2020b; Yue et al. 2020), uncertainty analysis (Zhang et al. 2020b, 2020c, 2020d), and so forth.

As Smarandache (1999) proposed neutrosophic sets (NSs) in 1999, Dai et al. (2011) developed the idea of SVNSs, which provided a better approach for dealing with uncertain and inaccurate information. Since then, multitudinous scholars have begun to launch intensive study on the theory and application of SVNSs. Özlü (2023a, b) and Özlü and Karaaslan (2022) developed single-valued neutrosophic type-2 hesitant fuzzy sets and gave some basic dice measures. Using the evaluation function, Mohamed et al. (2018) successfully combined TWD with SVNSs, and gave an AHP-QFD framework for resource selection. Ye (2015) optimized the cosine similarity measures of simplified NSs to realize the comparison among simplified NSs. On the basis of NSs, the representation of lattice combining TWD and fuzzy concept was discussed by Singh (2017). Furthermore, Singh (2018) proposed a lattice combining multi-granularity, TWD and n-valued neutrosophic concept. By summarizing various existing sorting methods, Huang et al. (2022) provided an approach for sorting the numbers in SVNSs (SVNNs).

In real-world data processing tasks, participants prefer to data with a hierarchical structure so as to obtain the knowledge of data in different hierarchies, then choose the data in a certain hierarchy that best suits their actual needs to proceed with processing. To address that above problem, Wu and Leung (2011, 2013) introduced the idea of multi-scale, that is, every object may have several values for a single attribute. Aiming at the issue of optimal scale selection (OPS) and reduction (OPR) in multi-scale decision systems (MSDSs), numerous scholars have launched discussions. In order to better identify and process the hyperspectral image, Dao et al. (2021) propose a novel OPS methodology. Li et al. (2017b) initially put forward the concept of the importance of multi-scale attributes and provided a simpler step-by-step OPS method. Wu et al. (2018a, b) considered the problem of global and local granularity selection in generalized decision systems with incomplete multi-granular labels and inconsistent generalized MSDSs respectively. Zhang et al. (2022) provided an updated way for OPS by using TWD and Hasse diagram.

Taking into account the illustrations above, the motives for this study are summarized below:

  1. (1)

    The existing researches on MSDSs are almost carried out under symbolic data environment, fuzzy data environment, intuitionistic fuzzy data environment and so on. Whereas there are few studies on MS-SVNDSs. For the sake of better solving problems with indeterminacy in real world, we establish the rough set model of MS-SVNDS.

  2. (2)

    According to the known research, the introduction of Pawlak rough sets, fuzzy rough sets, intuitionistic fuzzy rough sets and so on into MSDSs can effectively simplify data and improve the efficiency of rule extraction, so can the combination of single-valued neutrosophic rough sets and MSDSs perform better? For this problem, we discuss how to obtain optimal scales and optimal scale reducts using multi-scale single-valued neutrosophic rough sets and a pair of functions in evidence theory (DS).

  3. (3)

    The majority of previous studies focused on obtaining an optimal scale reduct of MSDSs and then make classification directly. While a good deal of judgments cannot be made immediately in reality, so it is necessary to introduce deferred decisions. Therefore, a novel TWD method on MS-SVNDSs is provided in this paper.

Comparison between this paper and our previous studies and the formation of this paper are displayed in Fig. 1. The organization of this paper is as follows. In the next section, we review the notions of TWD, SVNSs, MSDSs and DS. In Sect. 3, we construct a rough set model of MS-SVNDS and present the corresponding OPS and OPR methods. In Sect. 4, we give a novel TWD method. The experimental analyses of our method are carried out in Sect. 5. This paper's work is concluded in Sect. 6.

Fig. 1
figure 1

Comparison and formation of this paper

2 Preliminaries

We review the notions of TWD, SVNSs, MSDSs and DS in this section to facilitate the further analysis.

2.1 TWD

The following defines the TWD model as it was proposed by Yao (2012).

Definition 1

(Yao 2012). Assume \({O}^{\subseteq }\) is an object set which is non-empty and finite, suppose that \({v}_{a}:{O}^{\subseteq }\to {P}_{a}\) and \({v}_{r}:{O}^{\subseteq }\to {P}_{r}\) are respectively the acceptance evaluation function and the rejection evaluation function on \({O}^{\subseteq }\), where \(\left({P}_{a},{\le }_{a}\right)\) and \(\left({P}_{r},{\le }_{r}\right)\) are both partially ordered sets. Let \({P}_{a}^{+}\) and \({P}_{r}^{-}\) be non-empty subsets of \({P}_{a}\) and \({P}_{r}\). Then Three regions induced by \(\left({v}_{a},{v}_{r}\right)\) and \(\left({P}_{a}^{+},{P}_{r}^{-}\right)\) are defined as follows:

$$\begin{array}{c}{POS}_{\left({P}_{a}^{+},{P}_{r}^{-}\right)}\left({v}_{a},{v}_{r}\right)=\left\{x\in {O}^{\subseteq }|{v}_{a}\left(x\right)\in {P}_{a}^{+}\bigwedge {v}_{r}\left(x\right)\notin {P}_{r}^{-}\right\}\end{array}$$
(1)
$$\begin{array}{c}{NEG}_{\left({P}_{a}^{+},{P}_{r}^{-}\right)}\left({v}_{a},{v}_{r}\right)=\left\{x\in {O}^{\subseteq }|{v}_{a}\left(x\right)\notin {P}_{a}^{+}\bigwedge {v}_{r}\left(x\right)\in {P}_{r}^{-}\right\}\end{array}$$
(2)
$$\begin{array}{c}{BND}_{\left({P}_{a}^{+},{P}_{r}^{-}\right)}\left({v}_{a},{v}_{r}\right)=\left\{x\in {O}^{\subseteq }|\begin{array}{c}\left({v}_{a}\left(x\right)\in {P}_{a}^{+}\bigwedge {v}_{r}\left(x\right)\in {P}_{r}^{-}\right)\vee \\ \left({v}_{a}\left(x\right)\notin {P}_{a}^{+}\bigwedge {v}_{r}\left(x\right)\notin {P}_{r}^{-}\right)\end{array}\right\}\end{array}$$
(3)

Considering that it is unnecessary to use two evaluation functions in some practical situations, \({v}_{a}\) and \({v}_{r}\) can be combined into one function for evaluation.

Definition 2

(Yao 2012). Given \({O}^{\subseteq }\), suppose that \(v:{O}^{\subseteq }\to P\) is an evaluation function, where \(\left(P,\le \right)\) is a partially ordered set. Given two disjoint non-empty subsets \({P}^{+}\) and \({P}^{-}\) of \(P\), three regions induced by \(v\) and \(\left({P}^{+},{P}^{-}\right)\) are outlined below:

$$\begin{array}{c}{POS}_{\left({P}^{+},{P}^{-}\right)}\left(v\right)=\left\{x\in {O}^{\subseteq }|v\left(x\right)\in {P}^{+}\right\}\end{array}$$
(4)
$$\begin{array}{c}{NEG}_{\left({P}^{+},{P}^{-}\right)}\left(v\right)=\left\{x\in {O}^{\subseteq }|v\left(x\right)\in {P}^{-}\right\}\end{array}$$
(5)
$$\begin{array}{c}{BND}_{\left({P}^{+},{P}^{-}\right)}\left(v\right)=\left\{x\in {O}^{\subseteq }|v\left(x\right)\notin {P}^{+}\wedge v\left(x\right)\notin {P}^{-}\right\}\end{array}$$
(6)

Definition 3

(Yao 2012). Given \({O}^{\subseteq }\), \(\left(P,\le \right)\) is a totally ordered set, suppose that \(v:{O}^{\subseteq }\to P\) is an evaluation function. Given thresholds \(\alpha ,\beta \in P\) and \(\beta <\alpha\), three regions induced by \(v\) and \(\left(\alpha ,\beta \right)\) are outlined below:

$$\begin{array}{c}{POS}_{\left(\alpha ,\beta \right)}\left(v\right)=\left\{x\in {O}^{\subseteq }|v\left(x\right)\ge \alpha \right\}\end{array}$$
(7)
$$\begin{array}{c}{NEG}_{\left(\alpha ,\beta \right)}\left(v\right)=\left\{x\in {O}^{\subseteq }|v\left(x\right)\le \beta \right\}\end{array}$$
(8)
$$\begin{array}{c}{BND}_{\left(\alpha ,\beta \right)}\left(v\right)=\left\{x\in {O}^{\subseteq }|\beta <v\left(x\right)<\alpha \right\}\end{array}$$
(9)

2.2 SVNSs

To more accurately depict the ambiguity in real-world issues, Wang et al. (2010) provided the following SVNSs definition.

Definition 4

(Wang et al. 2010). Let \(U\) be an object set which is non-empty and finite, \(O=\left\{\langle x,{T}_{O}\left(x\right),{I}_{O}\left(x\right),{F}_{O}\left(x\right)\rangle |x\in U\right\}\) is a SVNS in \(U\), where \({T}_{O}\left(x\right)\in \left[\mathrm{0,1}\right]\) represents the degree of truth, \({I}_{O}\left(x\right)\in \left[\mathrm{0,1}\right]\) represents the degree of indeterminacy and \({F}_{O}\left(x\right)\in \left[\mathrm{0,1}\right]\) represents the degree of falsity. For \(\forall x\in U\), the SVNN in SVNS is noted as \(\left({T}_{O}\left(x\right),{I}_{O}\left(x\right),{F}_{O}\left(x\right)\right)\).

Shi and Ye (2017) further gave the definition of single-valued neutrosophic similarity degree.

Definition 5

(Shi and Ye 2017). Let \(U\) be an object set which is non-empty and finite, there are two SVNSs \({O}_{1}=\left\{\langle {x}_{i},{T}_{{O}_{1}}\left({x}_{i}\right),{I}_{{O}_{1}}\left({x}_{i}\right),{F}_{{O}_{1}}\left({x}_{i}\right)\rangle |{x}_{i}\in U, 1\le i\le n\right\}\) and \({O}_{2}=\left\{\langle {x}_{i},{T}_{{O}_{2}}\left({x}_{i}\right),{I}_{{O}_{2}}\left({x}_{i}\right),{F}_{{O}_{2}}\left({x}_{i}\right)\rangle \right.\left|\left.{x}_{i}\in U, 1\le i\le n\right\}\right.\) in \(U\), then.

$$\begin{array}{c}S\left({O}_{1},{O}_{2}\right)=\frac{1}{n}\sum_{i=1}^{n}\frac{{T}_{{O}_{1}}\left({x}_{i}\right){T}_{{O}_{2}}\left({x}_{i}\right)+{I}_{{O}_{1}}\left({x}_{i}\right){I}_{{O}_{2}}\left({x}_{i}\right)+{F}_{{O}_{1}}\left({x}_{i}\right){F}_{{O}_{2}}\left({x}_{i}\right)}{\sqrt{{T}_{{O}_{1}}^{2}\left({x}_{i}\right)+{I}_{{O}_{1}}^{2}\left({x}_{i}\right)+{F}_{{O}_{1}}^{2}\left({x}_{i}\right)}\sqrt{{T}_{{O}_{2}}^{2}\left({x}_{i}\right)+{I}_{{O}_{2}}^{2}\left({x}_{i}\right)+{F}_{{O}_{2}}^{2}\left({x}_{i}\right)}}\end{array}$$
(10)

is defined as the similarity degree between \({O}_{1}\) and \({O}_{2}\).

Property 1

(Shi and Ye 2017). Given two SVNSs \({O}_{1}=\left\{\langle x,{T}_{{O}_{1}}\left(x\right),{I}_{{O}_{1}}\left(x\right),{F}_{{O}_{1}}\left(x\right)\rangle |x\in U\right\}\) and \({O}_{2}=\left\{\langle x,{T}_{{O}_{2}}\left(x\right),{I}_{{O}_{2}}\left(x\right),{F}_{{O}_{2}}\left(x\right)\rangle |x\in U\right\}\), \(S\left({O}_{1},{O}_{2}\right)\) is the similarity degree between \({O}_{1}\) and \({O}_{2}\), then

  1. (1)

    \(\left(1\right)0\le S\left({O}_{1},{O}_{2}\right)\le 1;\)

  2. (2)

    \(\left(2\right)S\left({O}_{1},{O}_{2}\right)=S\left({O}_{2},{O}_{1}\right);\)

  3. (3)

    If \(S\left({O}_{1},{O}_{2}\right)=1\), then \({T}_{{O}_{1}}\left(x\right)={T}_{{O}_{2}}\left(x\right)\), \({I}_{{O}_{1}}\left(x\right)\)=\({I}_{{O}_{2}}\left(x\right)\), \({F}_{{O}_{1}}\left(x\right)={F}_{{O}_{2}}\left(x\right).\)

2.3 MSDSs

Definition 6

(Wu and Leung 2011). Tuple \(S=\left(U,AT,V,f\right)\) represents a multi-scale information system, where \(U=\left\{{x}_{1},{x}_{2},\dots ,{x}_{n}\right\}\) is the universe of discourse, \(AT=\left\{{a}_{1},{a}_{2},\dots ,{a}_{m}\right\}\) is a non-empty and finite set of conditional attributes, \(\forall {a}_{j}\in AT \left(1\le j\le m\right)\) can assume various values at various scales. \(V\) is the domain of conditional attributes, \(f\) is a map from \(U\times AT\) onto \(V\). \(S=\left(U,AT\bigcup \left\{d\right\},V,f\right)\) is considered to be a MSDS, where \(d\notin AT\) is a decision attribute.

Above is the definition of the original Wu–Leung model (2011, 2013)). Based on it, Li and Hu (2017) provided a novel model definition.

Definition 7

(Li and Hu 2017). Let \(S=\left(U,AT,V,f\right)=\left(U,\left\{{a}_{j}^{k}|k=\mathrm{1,2},\dots ,{I}_{j}, j=\mathrm{1,2},\dots ,m\right\},V,f\right)\) be a multi-scale information system, where \({a}_{j}^{k}:U\to {V}_{j}^{k}\) is a surjective mapping and \({V}_{j}^{k}\) is the domain of the attribute \({a}_{j}\) on the \(k\) th scale. For \(\forall 1\le k\le {I}_{j}-1\), there exists a surjective mapping \({g}_{j}^{k,k+1}:{V}_{j}^{k}\to {V}_{j}^{k+1}\) such that \(a_{j}^{{k + 1}} = g_{j}^{{{k,k + 1}} ^\circ} a_{j}^{k}\), \({a}_{j}^{k+1}\left(x\right)={g}_{j}^{k,k+1}\left({a}_{j}^{k}\left(x\right)\right)\), \(x\in U\). Based on the above, we consider \(S=\left(U,AT\bigcup \left\{d\right\},V,f\right)\) a MSDS, where \(d\notin AT\) is a decision attribute.

Definition 8

(Li and Hu 2017). Let \(S=\left(U,AT\bigcup \left\{d\right\},V,f\right)\) be a MSDS, where each attribute \({a}_{j}\) has \({I}_{j}\) scales \(\left(1\le j\le m\right)\). There forms a decision system \({S}^{K}\) when attributes \({a}_{1},{a}_{2},\dots ,{a}_{m}\) are respectively restricted on their \({l}_{j}\) th scale \(\left(1\le j\le m\right)\), where \(K=\left({l}_{1},{l}_{2},\dots ,{l}_{m}\right)\) is the scale combination of \({S}^{K}\) in \(S\). \(\mathcal{L}=\left\{\left({l}_{1},{l}_{2},\dots ,{l}_{m}\right)|1\le {l}_{j}\le {I}_{j},j=\mathrm{1,2},\dots ,m\right\}\) is the scale collection of \(S\), which contains all the scale combinations in \(S\).

2.4 DS

For better coping with uncertain information and uncertain situations, DS was put forward.

Definition 9

(Niu et al. 2019). Let \({S}^{K}=\left(U,{AT}^{K}\bigcup \left\{d\right\},V,f\right)\), \(X\subseteq U\), \({R}_{{AT}^{K}}\) is the equivalent relation on \({S}^{K}\), \(\underline{{R_{{AT^{K} }} }} \left( X \right)\) and \(\overline{{R}_{{AT}^{K}}}\left(X\right)\) are respectively the lower approximation and upper approximation of \(X\) w.r.t. \({R}_{{AT}^{K}}\). Then the belief function and plausibility function of \(U\) are respectively defined as

$$Bel_{{AT^{K} }} \left( X \right) = P\left( {\underline{{R_{{AT^{K} }} }} \left( X \right)} \right) = \frac{{\left| {\underline{{R_{{AT^{K} }} }} \left( X \right)} \right|}}{{\left| U \right|}}$$
(11)
$$Pl_{{AT^{K} }} \left( X \right) = P\left( {\overline{{R_{{AT^{K} }} }} \left( X \right)} \right) = \frac{{\left| {\overline{{R_{{AT^{K} }} }} \left( X \right)} \right|}}{{\left| U \right|}}$$
(12)

where the corresponding mass function is noted as

$$m_{{AT^{K} }} \left( X \right) = \left\{ {\begin{array}{ll} {P\left( X \right) = \frac{{\left| X \right|}}{{\left| U \right|}},} & {X \in \frac{U}{{R_{{AT^{K} }} }}} \\ {0,} & {X \notin \frac{U}{{R_{{AT^{K} }} }}} \\ \end{array} } \right.$$
(13)

Hence for \({D}_{i}\in U/d,\) there exists

$$\begin{array}{c}{Bel}_{{AT}^{K}}\left(d\right)=\sum_{{D}_{i}\in \frac{U}{d}}{Bel}_{{AT}^{K}}\left({D}_{i}\right)=\sum_{{D}_{i}\in \frac{U}{d}}\frac{\left|\underset{\_}{{R}_{{AT}^{K}}}\left({D}_{i}\right)\right|}{\left|U\right|}\end{array}$$
(14)
$$\begin{array}{c}{Pl}_{{AT}^{K}}\left(d\right)=\sum_{{D}_{i}\in \frac{U}{d}}{Pl}_{{AT}^{K}}\left({D}_{i}\right)=\sum_{{D}_{i}\in \frac{U}{d}}\frac{\left|\overline{{R}_{{AT}^{K}}}\left({D}_{i}\right)\right|}{\left|U\right|}\end{array}$$
(15)

3 OPS and OPR of MS-SVNDS

We give the definition of the rough set model of MS-SVNDS in this section. The corresponding OPS and OPR methods are further explored. Hence, we have laid the environmental foundation for the discussion of the following TWD method on MS-SVNDS.

3.1 MS-SVNDS

First, we construct the rough set model of MS-SVNDS and give the related properties.

Definition 10

Let \(MS-SVNDS=\left(U,AT\bigcup \left\{d\right\},V,f\right)\) be a multi-scale single-valued neutrosophic decision table, where \(U=\left\{{x}_{1},{x}_{2},\dots ,{x}_{n}\right\}\) is the universe of discourse, \(AT=\left\{{a}_{1},{a}_{2},\dots ,{a}_{m}\right\}\) denotes a non-empty finite set of conditional attributes and each \({a}_{j}\in AT\) has \({I}_{j}\) scales \(\left(1\le j\le m\right)\). When \({a}_{1},{a}_{2},\dots ,{a}_{m}\) are respectively restricted on their \({l}_{j}\) th \(\left(1\le j\le m\right)\) scale, there form a single-valued neutrosophic decision table \({S}^{K}\), where \(K=\left({l}_{1},{l}_{2},\dots ,{l}_{m}\right) \left(1\le {l}_{j}\le {I}_{j},1\le j\le m\right)\in \mathcal{L}\) is denoted as the scale combination of \({S}^{K}\) in \(S\) and \(\mathcal{L}=\left\{\left({l}_{1},{l}_{2},\dots ,{l}_{m}\right)|1\le {l}_{j}\le {I}_{j},j=\mathrm{1,2},\dots ,m\right\}\) is the scale collection of \(S\). \(d\notin AT\) is a decision attribute, \(U/d=\left\{{D}_{1},{D}_{2},\dots ,{D}_{m}\right\}\) constitutes an accurate division of \(U\). \(V\) is the domain of \(AT\bigcup \left\{d\right\}\) and \(f\): \(U\times \left(AT\bigcup \left\{d\right\}\right)\) \(\to\) \(V\) is a mapping function. For \(\forall x\in U, a\in AT, f\left(x,a\right)=\left({T}_{x}\left(a\right),{I}_{x}\left(a\right),{F}_{x}\left(a\right)\right) \left({T}_{x}\left(a\right)\in \left[\mathrm{0,1}\right],{I}_{x}\left(a\right)\in \left[\mathrm{0,1}\right],{F}_{x}\left(a\right)\in \left[\mathrm{0,1}\right]\right)\) represents the SVNN of object \(x\) under the conditional attribute \(a\). The finest scale combination \({K}_{0}=\left({l}_{1}=1,{l}_{2}=1,\dots ,{l}_{m}=1\right)\). For \(\forall {a}_{j}\in AT \left(1\le j\le m\right)\), we have \({a}_{j}^{1}\left({x}_{i}\right)\le {a}_{j}^{2}\left({x}_{i}\right)\le \dots \le {a}_{j}^{{I}_{j}}\left({x}_{i}\right) \left(1\le i\le n,1\le j\le m\right)\).

Definition 11

Let \(MS-SVNDS=\left(U,AT\bigcup \left\{d\right\},V,f\right)\), where \(AT=\left\{{a}_{1},{a}_{2},\dots ,{a}_{m}\right\}\) and each \({a}_{j}\in AT\) has \({I}_{j}\) scales \(\left(1\le j\le m\right)\), scale combination \(K=\left({l}_{1},{l}_{2},\dots ,{l}_{m}\right) \left(1\le {l}_{j}\le {I}_{j},1\le j\le m\right)\), for \(\forall x\in U, a\in AT, f\left(x,a\right)=\left({T}_{x}\left(a\right),{I}_{x}\left(a\right),{F}_{x}\left(a\right)\right) \left({T}_{x}\left(a\right)\in \left[\mathrm{0,1}\right],{I}_{x}\left(a\right)\in \left[\mathrm{0,1}\right],{F}_{x}\left(a\right)\in \left[\mathrm{0,1}\right]\right)\), \({D}_{i}\subseteq U\left(1\le i\le m\right)\),

$$\begin{array}{c}{S}^{K}\left(x,y\right)=\frac{1}{m}\sum_{j=1}^{m}\frac{{T}_{x}\left({a}_{j}^{{l}_{j}}\right){T}_{y}\left({a}_{j}^{{l}_{j}}\right)+{I}_{x}\left({a}_{j}^{{l}_{j}}\right){I}_{y}\left({a}_{j}^{{l}_{j}}\right)+{F}_{x}\left({a}_{j}^{{l}_{j}}\right){F}_{y}\left({a}_{j}^{{l}_{j}}\right)}{\sqrt{{T}_{x}^{2}\left({a}_{j}^{{l}_{j}}\right)+{I}_{x}^{2}\left({a}_{j}^{{l}_{j}}\right)+{F}_{x}^{2}\left({a}_{j}^{{l}_{j}}\right)}\sqrt{{T}_{y}^{2}\left({a}_{j}^{{l}_{j}}\right)+{I}_{y}^{2}\left({a}_{j}^{{l}_{j}}\right)+{F}_{y}^{2}\left({a}_{j}^{{l}_{j}}\right)}}\end{array}$$
(16)
$$1\le j\le m, 1\le {l}_{j}\le {I}_{j}$$

is denoted as the single-valued neutrosophic similarity degree between object \(x\) and object \(y\) in \(U\) on \(K\). Then, \({S}^{K}\) is the single-valued neutrosophic similarity relation induced by \(K\).

$$\begin{array}{c}{\left[x\right]}_{{S}^{K}}=\sum_{y\in U}\frac{{S}^{K}\left(x,y\right)}{y}\end{array}$$
(17)

is the single-valued neutrosophic similarity set w.r.t. \(x\) derived from \(K\).

$$\begin{array}{c}\underset{\_}{{S}^{K}}\left({D}_{i}\right)\left(x\right)=\left\{x\in U\left|{\left[x\right]}_{{S}^{K}}\subseteq {D}_{i}\right.\right\}\end{array}$$
(18)
$$\begin{array}{c}\overline{{S}^{K}}\left({D}_{i}\right)\left(x\right)=\left\{x\in U\left|{\left[x\right]}_{{S}^{K}}\cap {D}_{i}\ne \varnothing \right.\right\}\end{array}$$
(19)

We call the pairs \(\left(\underset{\_}{{S}^{K}}\left({D}_{i}\right),\overline{{S}^{K}}\left({D}_{i}\right)\right)\) as rough sets of \({D}_{i}\) w.r.t. \(K\) in \(MS-SVNDS\). If \(\underset{\_}{{S}^{K}}\left({D}_{i}\right)=\overline{{S}^{K}}\left({D}_{i}\right)\), then \({D}_{i}\) is called definable w.r.t. \(K\); if not, \({D}_{i}\) is called undefinable w.r.t. \(K\).

For \({D}_{i}\in U/d\), the belief function and plausibility function w.r.t. \(K\) are defined below:

$$\begin{array}{c}{Bel}_{K}\left(d\right)=\sum_{{D}_{i}\in \frac{U}{d}}{Bel}_{K}\left({D}_{i}\right)=\sum_{{D}_{i}\in \frac{U}{d}}\frac{\left|\underset{\_}{{S}^{K}}\left({D}_{i}\right)\right|}{\left|U\right|}\end{array}$$
(20)
$$\begin{array}{c}{Pl}_{K}\left(d\right)=\sum_{{D}_{i}\in \frac{U}{d}}{Pl}_{K}\left({D}_{i}\right)=\sum_{{D}_{i}\in \frac{U}{d}}\frac{\left|\overline{{S}^{K}}\left({D}_{i}\right)\right|}{\left|U\right|}\end{array}$$
(21)

Proposition 1

The following properties hold based on Definition 9:

  1. (1)

    \(\forall {D}_{i}\subseteq U\), \(\underline{{S^{K} }} \left( {D_{i} } \right) \subseteq \overline{{S^{K} }} \left( {D_{i} } \right);\)

  2. (2)

    \(\underline{{S^{K} }} \left( U \right) = \overline{{S^{K} }} \left( U \right) = U;\)

3.2 OPS of MS-SVNDS

On the basis of the exposition in the previous section, each object has a set of attributes with several scales and each attribute takes on different SVNNs at different scale levels when in a MS-SVNDS. Obviously, the workload of directly gaining the rules from such a huge amount of data is very difficult and complicated. Therefore, in order to effectively solving the problem of knowledge acquisition, we discuss the OPS method of MS-SVNDS in this section.

For the purpose of elaborating the OPS method of MS-SVNDS, we first review the lattice model proposed by Li and Hu (Li and Hu 2017) as following.

Definition 12

(Li and Hu 2017). Let \({K}_{1}=\left({l}_{1}^{1},{l}_{2}^{1},\dots ,{l}_{m}^{1}\right)\), \({K}_{2}=\left({l}_{1}^{2},{l}_{2}^{2},\dots ,{l}_{m}^{2}\right)\in \mathcal{L}\).

$$\begin{array}{c}{K}_{1}\le {K}_{2}\iff {l}_{{\text{j}}}^{1}\le {l}_{{\text{j}}}^{2}, 1\le j\le m\end{array}$$
(22)
$$\begin{array}{c}{K}_{1}={K}_{2}\iff {l}_{{\text{j}}}^{1}={l}_{{\text{j}}}^{2}, 1\le j\le m\end{array}$$
(23)
$$\begin{array}{c}{K}_{1}<{K}_{2}\iff {K}_{1}\le {K}_{2}\wedge {K}_{1}\ne {K}_{2}\end{array}$$
(24)
$$\begin{array}{c}{K}_{1}\wedge {K}_{2}=\left(min\left({l}_{1}^{1},{l}_{1}^{2}\right),min\left({l}_{2}^{1},{l}_{2}^{2}\right),\cdots ,min\left({l}_{m}^{1},{l}_{m}^{2}\right)\right)\end{array}$$
(25)
$$\begin{array}{c}{K}_{1}\vee {K}_{2}=\left(max\left({l}_{1}^{1},{l}_{1}^{2}\right),max\left({l}_{2}^{1},{l}_{2}^{2}\right),\cdots ,max\left({l}_{m}^{1},{l}_{m}^{2}\right)\right)\end{array}$$
(26)

Perceptibly, \(\left(1, 1, \cdots , 1\right)\le \left({l}_{1},{l}_{2},\dots ,{l}_{m}\right)\le \left({I}_{1},{I}_{2}, \cdots , {I}_{m}\right)\).

\(\left(\mathcal{L},\le ,\wedge ,\vee \right)\) is referred to as a lattice, where \(\mathcal{L}\) is the scale collection of MS-SVNDS, \(\le\) is a partial relation, \(\wedge\) and \(\vee\) respectively stand for the operation of taking the maximum and the minimum value.

The definitions of optimal scale combinations of MS-SVNDS are given below.

Definition 13

Let \(MS-SVNDS=\left(U,AT\bigcup \left\{d\right\},V,f\right)\), where \(AT=\left\{{a}_{1},{a}_{2},\dots ,{a}_{m}\right\}\) and \(d\notin AT\) constitutes a set of divisions \(U/d=\left\{{D}_{1},{D}_{2},\dots ,{D}_{m}\right\}\). \(K=\left({l}_{1},{l}_{2},\dots ,{l}_{m}\right)\) and \({K}_{0}=\left(1, 1,\dots , 1\right)\) are respectively a combination in \(\mathcal{L}\) and the finest combination in \(\mathcal{L}\), \(\mathcal{L}=\left\{\left({l}_{1},{l}_{2},\dots ,{l}_{m}\right)|1\le {l}_{j}\le {I}_{j},j=\mathrm{1,2},\dots ,m\right\}\).

  1. (1)

    If there exists \({Bel}_{K}\left(d\right)={Bel}_{{K}_{0}}\left(d\right)\), \(K\) is then considered as a consistent belief scale. Based on above, if there always exists \({Bel}_{{K}{\prime}}\left(d\right)\ne {Bel}_{{K}_{0}}\left(d\right)\) for \(\forall {K}{\prime}\in \mathcal{L}\) that satisfies \({K}{\prime}>K\), \(K\) is then considered as an optimal belief scale.

  2. (2)

    If there exists \({Pl}_{K}\left(d\right)={Pl}_{{K}_{0}}\left(d\right)\), \(K\) is then considered as a consistent plausibility scale. Based on above, if there always exists \({Pl}_{{K}{\prime}}\left(d\right)\ne {Pl}_{{K}_{0}}\left(d\right)\) for \(\forall {K}{\prime}\in \mathcal{L}\) that satisfies \({K}{\prime}>K\), \(K\) is then considered as an optimal plausibility scale.

In view of the above, we give the OPS algorithm of MS-SVNDS in the following.

figure a

Algorithm 1 Optimal belief scale selection algorithm of MS-SVNDS

In the above algorithm, we first initialize \(\varphi\) to be equal to the scale collection \(\mathcal{L}\). Next, we separately compute the belief functions w.r.t. \({K}_{0}\) and \(K \left(\forall K\in \mathcal{L}\right)\). By comparing the belief functions w.r.t. the above two, we determine whether to delete or retain \(K\), i.e. delete \(K\) whose belief function is different from the belief function w.r.t. \({K}_{0}\). Then, we delete all scales finer than the reserved scales and obtain the optimal belief scales of MS-SVNDS.

One can achieve the optimal plausibility scale selection procedure of MS-SVNDS by replacing \({Bel}_{{K}_{0}}\left(d\right)\) with \({Pl}_{{K}_{0}}\left(d\right)\) and replacing \({Bel}_{K}\left(d\right)\) with \({Pl}_{K}\left(d\right)\). Apparently, Algorithm 1's time complexity is \(O\left({n}^{2}\prod_{j=1}^{m}{I}_{j}\right)\).

3.3 OPR of MS-SVNDS

After selecting the optimal scales of MS-SVNDS, we acquire the coarsest scales that are coordinated with the finest scale, which effectively simplifies the data. However, not all attributes in MS-SVNDS are necessary and some of them may lead to redundancy. Thus, we discuss the acquisition of optimal scale reducts of MS-SVNDS in this section.

We start by providing the following description of sub-scale. Then, the OPR approach of MS-SVNDS is subsequently presented.

Definition 14

Let\(MS-SVNDS=\left(U,AT\bigcup \left\{d\right\},V,f\right)\), \(K=\left({l}_{1},{l}_{2},\dots ,{l}_{m}\right) \in \mathcal{L}\), sub-scale \({K}^{\subseteq }=\left(-,\cdots ,-, {l}_{i}, -,\dots ,-,{l}_{j},-,\cdots ,-\right)\subseteq K,\) \(\mathcal{L}\) is the scale collection of MS-SVNDS. When there exists\({K}^{\subseteq }\subset K\), \({K}^{\subseteq }\) is said to be a proper sub-scale, where "-" denotes the deletion of the relevant attribute.

Definition 15

Let \(MS-SVNDS=\left(U,AT\bigcup \left\{d\right\},V,f\right)\), \(K=\left({l}_{1},{l}_{2},\dots ,{l}_{m}\right) \in \mathcal{L}\), \({K}^{\subseteq }\subseteq K\), \(U/d=\left\{{D}_{1},{D}_{2},\dots ,{D}_{m}\right\}\).

  1. (1)

    If \({Bel}_{{K}^{\subseteq }}\left(d\right)={Bel}_{K}\left(d\right)\) is met, then \({K}^{\subseteq }\) is called as a consistent belief sub-scale of \(K\). When \({K}^{\subseteq }\) is the only one that meets the above conditions, \({K}^{\subseteq }\) is referred to as a consistent belief scale reduct of \(K\).

  2. (2)

    If \({Pl}_{{K}^{\subseteq }}\left(d\right)={Pl}_{K}\left(d\right)\) is met, then \({K}^{\subseteq }\) is called as a consistent plausibility sub-scale of \(K\). When \({K}^{\subseteq }\) is the only one that meets the above conditions, \({K}^{\subseteq }\) is referred to as a consistent plausibility scale reduct of \(K\).

Evidently, we can get the reducts of all optimal scales according to the above method. In order to handle the circumstance that we need to quickly obtain a set of the most refined knowledge in real life more effectively, we provide an algorithm to directly obtain an optimal scale reduct of MS-SVNDS.

figure b

Algorithm 2 Obtaining an optimal belief scale reduct of MS-SVNDS

In Algorithm 2, we first initialize \({AT}{\prime}\) to be equal to the conditional attribute set \(AT\) and \({K}{\prime}\) to be equal to the finest scale \({K}_{0}\). Next, we separately compute the belief functions w.r.t. \({K}{\prime}\) and \({K}{\prime}\backslash \left\{{a}_{j}\right\} \left(\forall K\in \mathcal{L}\right)\). We delete \({a}_{j}\) when the belief function w.r.t. \({K}{\prime}\backslash \left\{{a}_{j}\right\}\) are the same as the belief function w.r.t. \({K}{\prime}\). Otherwise, we raise the scale level of the reserved \({a}_{j}\) to the maximum under the condition of keeping the belief function unchanged. After checking the conditional attributes one by one and outputting the final \({K}{\prime}\), we acquire an optimal belief scale reduct of MS-SVNDS.

The algorithm for calculating an optimal plausibility scale reduct of MS-SVNDS can be obtained by replacing all belief functions with plausibility functions. Apparently, Algorithm 2's time complexity is is \(O\left({n}^{2}\prod_{j=1}^{m}{I}_{j}\right)\).

4 A TWD method on MS-SVNDS

In the previous section, we construct the rough set model of MS-SVNDS, the selection and reduction of the optimal scale combinations are also investigated. Thus the effective knowledge in the decision table can be obtained and reliable decisions can be made. However, considering that not all decision problems in reality can get two diametrically opposed results, we introduce TWD method to classify objects into three different categories: non-commitment, rejection and acceptance, thereby forming more accurate decision results.

Table 1 Datasets details

In the contents that follow, we address the TWD issue directly based on the optimal scale reduct obtained by Algorithm 2.

Definition 16

Let \(OP-SVNDS=\left(U,AT\bigcup \left\{d\right\},V,f\right)\), where \(U=\left\{{x}_{1},{x}_{2},\dots ,{x}_{n}\right\}\) and \(AT=\left\{-,\cdots ,-, {a}_{i}^{{l}_{i}}, -,\dots ,-,{a}_{j}^{{l}_{j}},-,\cdots ,-\right\}\left(1\le i\le j\le m, 1\le {l}_{i}, {l}_{j}\le {I}_{j}\right)\), each conditional attribute in \(AT\) is restricted on its local optimal scale level and “\(-\)” represents that the corresponding conditional attribute is deleted. For \(\forall x\in U, a\in AT, f\left(x,a\right)=\left({T}_{x}\left(a\right),{I}_{x}\left(a\right),{F}_{x}\left(a\right)\right) \left({T}_{x}\left(a\right)\in \left[\mathrm{0,1}\right],{I}_{x}\left(a\right)\in \left[\mathrm{0,1}\right],{F}_{x}\left(a\right)\in \left[\mathrm{0,1}\right]\right)\). \(OP=\left\{-,\cdots ,-, {l}_{i}, -,\dots ,-,{l}_{j},-,\cdots ,-\right\}\) is the scale combination of \(OP-SVNDS\).

$$\begin{array}{c}{S}^{OP}=\frac{1}{n}\sum_{j=1}^{m}\frac{{T}_{x}\left({a}_{j}^{{l}_{j}}\right){T}_{y}\left({a}_{j}^{{l}_{j}}\right)+{I}_{x}\left({a}_{j}^{{l}_{j}}\right){I}_{y}\left({a}_{j}^{{l}_{j}}\right)+{F}_{x}\left({a}_{j}^{{l}_{j}}\right){F}_{y}\left({a}_{j}^{{l}_{j}}\right)}{\sqrt{{T}_{x}^{2}\left({a}_{j}^{{l}_{j}}\right)+{I}_{x}^{2}\left({a}_{j}^{{l}_{j}}\right)+{F}_{x}^{2}\left({a}_{j}^{{l}_{j}}\right)}\sqrt{{T}_{y}^{2}\left({a}_{j}^{{l}_{j}}\right)+{I}_{y}^{2}\left({a}_{j}^{{l}_{j}}\right)+{F}_{y}^{2}\left({a}_{j}^{{l}_{j}}\right)}}\end{array}$$
(27)
$${a}_{j}\ne -, 1\le j\le m, 1\le {l}_{j}\le {I}_{j}$$

is denoted as the single-valued neutrosophic similarity function on \(OP\), where \(n\in \left(0,\right.\left.m\right]\) is the number of reserved conditional attributes in \(AT\).

$$\begin{array}{c}{\left[x\right]}_{{S}_{\omega }^{OP}}=\left\{y\in U|{S}^{OP}\left(x,y\right)\ge \omega \right\}\end{array}$$
(28)

is the \(\omega\)-level (\(\omega \in \left(0,\right.\left.1\right]\)) single-valued neutrosophic similarity class w.r.t. \(x\) derived from \(OP\), where \({S}_{\omega }^{OP}\) is called a \(\omega\)-level single-valued neutrosophic similarity relation.

Definition 17

Let \(OP-SVNDS=\left(U,AT\bigcup \left\{d\right\},V,f\right)\) and \(\varphi ({O}_{1},{O}_{2})=\frac{\left|{O}_{1}\cap {O}_{2}\right|}{\left|{O}_{1}\right|}\) be an evaluation function. Given thresholds \(\alpha , \beta \in \left(0,\right.\left.1\right]\) and \(\alpha >\beta\). For any \(O\subseteq U\),

$$\begin{array}{c}{POS}_{\left(\alpha , \beta \right)}\left(O\right)=\left\{x\in U|\varphi \left({\left[x\right]}_{{S}_{\omega }^{OP}},O\right)\ge \alpha \right\}\end{array}$$
(29)
$$\begin{array}{c}{NEG}_{\left(\alpha , \beta \right)}\left(O\right)=\left\{x\in U|\varphi \left({\left[x\right]}_{{S}_{\omega }^{OP}},O\right)\le \beta \right\}\end{array}$$
(30)
$$\begin{array}{c}{BND}_{\left(\alpha , \beta \right)}\left(O\right)=\left\{x\in U|\beta <\varphi \left({\left[x\right]}_{{S}_{\omega }^{OP}},O\right)<\alpha \right\}\end{array}$$
(31)

Proposition 2

Given a three-way decision model \(\left(U,{S}_{\omega }^{OP},\varphi ,\alpha ,\beta \right)\), for \({O}_{1},{O}_{2},{O}_{3}\subseteq U and {O}_{1}\subseteq {O}_{2}\), if there exists \(\varphi \left({O}_{3},{O}_{1}\right)\le \varphi \left({O}_{3},{O}_{2}\right)\), then

  1. (1)

    \({POS}_{\left(\alpha , \beta \right)}\left({O}_{1}\right)\subseteq {POS}_{\left(\alpha , \beta \right)}\left({O}_{2}\right);\)

  2. (2)

    \({NEG}_{\left(\alpha , \beta \right)}\left({O}_{2}\right)\subseteq {NEG}_{\left(\alpha , \beta \right)}\left({O}_{1}\right).\)

Proof

(1) Obviously, for \(\forall x\in {POS}_{\left(\alpha , \beta \right)}\left({O}_{1}\right)\), there is \(\varphi \left({\left[x\right]}_{{S}_{\omega }^{OP}},{O}_{1}\right)\ge \alpha\). Considering that \(\varphi \left({\left[x\right]}_{{S}_{\omega }^{OP}},{O}_{2}\right)\ge \varphi \left({\left[x\right]}_{{S}_{\omega }^{OP}},{O}_{1}\right)\), we have \(\varphi \left({\left[x\right]}_{{S}_{\omega }^{OP}},{O}_{2}\right)\ge \alpha\), which represents that \(x\in {POS}_{\left(\alpha , \beta \right)}\left({O}_{2}\right)\). Therefore, we obtain that \({POS}_{\left(\alpha , \beta \right)}\left({O}_{1}\right)\subseteq {POS}_{\left(\alpha , \beta \right)}\left({O}_{2}\right)\).

(2) For \(\forall x\in {NEG}_{\left(\alpha , \beta \right)}\left({O}_{2}\right)\), there is \(\varphi \left({\left[x\right]}_{{S}_{\omega }^{OP}},{O}_{2}\right)\le \beta\). According to \(\varphi \left({\left[x\right]}_{{S}_{\omega }^{OP}},{O}_{1}\right)\le \varphi \left({\left[x\right]}_{{S}_{\omega }^{OP}},{O}_{2}\right)\), we have \(\varphi \left({\left[x\right]}_{{S}_{\omega }^{OP}},{O}_{1}\right)\le \beta\), which represents that \(x\in {NEG}_{\left(\alpha , \beta \right)}\left({O}_{1}\right)\). Therefore, we obtain that \({NEG}_{\left(\alpha , \beta \right)}\left({O}_{2}\right)\subseteq {NEG}_{\left(\alpha , \beta \right)}\left({O}_{1}\right)\).

On an optimal scale reduct of MS-SVNDS, we present the specific algorithm for the TWD approach as follows.

figure c

Algorithm 3 The algorithm of TWD method

5 Experimental analysis

In this section, we employ experimental analysis to confirm the accuracy and viability of our method. Figure 2 shows the experimental process. The experiment was conducted with MATLAB R2022a. We utilize seven UCI data sets to assess the efficacy of our method, i.e. Wine, Iris, Algerian Forest Fires, Glass, Thyroid, Bupa and Seeds. Table 1 shows these data sets' specifics.

Fig. 2
figure 2

Experimental flow chart

All data sets are preprocessed before the evaluation experiment. Since the seven data sets are all single-scale, the multi-scale environments of them are firstly constructed. Use the data set Thyroid as an illustration. Figure 3 and 4 present the feature distribution and visualization of Thyroid. Based on the approach in Wu and Leung (2011), we first find out the maximum and minimum values of each feature in Thyroid and do subtraction between the two. Then the difference value obtained from each feature can be divided by 4 to get a 4-value interval \(\left[\frac{i-1}{4}\left({A}_{max}-{A}_{min}\right),\frac{i}{4}\left({A}_{max}-{A}_{min}\right)\right]\) as the value on the coarser scale level, and so on to construct multiple scale hierarchies. After the above steps, the scale levels of features in Thyroid are finally set to (1,3,2,2,3).

After constructing the multi-scale environment, we then convert all the data in data set into SVNNs. We start by calculating the Euclidean distance between each feature value and the maximum, minimum, and average values in the feature column, then use the absolute value that results from deducting the normalized distance from 1 to determine the degree to which the feature in the data set is true, indeterminate or false. Thus, we have established a rough set model of MS-SVNDS in the data set.

Fig. 3
figure 3

Feature distribution of Thyroid

Fig. 4
figure 4

Visualization of Thyroid

Fig. 5
figure 5

Feature distribution of Thyroid after reduction

Fig. 6
figure 6

Visualization of Thyroid after reduction

Fig. 7
figure 7

Comparison of number of features

Fig. 8
figure 8

Comparison of classification accuracy

In the next step, we employ our Algorithm 2 to acquire an optimal scale reduct based on the established model. First, we calculate the belief function when all the features are at the first scale level, and then eliminate each feature one at a time to determine whether the belief function following the removal of a particular feature is equivalent to the belief function before. If so, confirm the deletion; If not, the belief function on the coarse scale of the feature is examined to see if it equals the original belief function. If they are the same, the scale hierarchy of the feature will be upgraded accordingly; if they are not, the original scale hierarchy of the feature will be kept. The final scale level of the feature is the maximum value that satisfies the above equality condition. Thus, we can get an optimal scale reduct of the data set, which simplifies the data to some extent.

Still take Thyroid as an example. The reserved features of Thyroid after OPR and their corresponding scale levels are (1,2,2,2,0), in which the feature of ‘0’ represents deleted. The feature distribution and visualization of Thyroid after processing are shown in Fig. 5 and 6.

As shown in Table 2, we construct a multi-scale environment in each data set. By using our algorithm to process the multi-scale data sets, several redundant features are deleted and the scale level of some features is improved, thus effectively simplifying the data. The average classification accuracy of the comparison algorithm in Wang et al. (2023) and our algorithm obtained by 5 times of tenfold cross validation with KNN algorithm is shown in Table 3. Apparently, our algorithm can obtain relatively high average classification accuracy with simplified data sets, and the classification performance of our algorithm is improved compared with the comparison algorithm, which verifies the effectiveness and feasibility of our algorithm. The comparison of the experimental results is shown in Fig. 7 and 8.

Table 2 OPS and OPR results
Table 3 Classification results

To carry out further tests, we employ the win/draw/loss analysis presented by Chen et al. (2020) for sign and binomial test of classification accuracy. A total of 50 test samples of classification accuracy are obtained by performing 5 times of tenfold cross validation using the comparison algorithm and our algorithm. The \(p\) values here are the two-tail sign and binomial test results. It can be clearly concluded from Table 4 that our algorithm can achieve better classification effects.

Table 4 Statistical test results

Upon the simplification of the data set, we recalculate the similarity degree between the objects and divide them into ω-level similarity classes by setting the similarity threshold ω. We finally get the partition of objects by comparing the evaluation values of ω-level similarity classes and subsets of objects with the introduced thresholds \(\alpha\) and \(\beta\). The three-way decision results of data set Thyroid are shown below, where \(\omega\) is set to 0.7259.

Table 5 Classification results of all alternatives in the data set ‘Thyroid’

The processed data set Thyroid consists of 215 alternatives and 4 features. The scale levels of 4 features are respectively 1, 2, 2, 2. From the classification results of Thyroid in Table 5, we can find that 35 alternatives are considered to belong to \(POS(F)\), 34 alternatives are considered to belong to \(BND(F)\) and 146 alternatives are considered to belong to \(NEG(F)\) in view of Class 2. It means that it can be confirmed that Class 2 contains at least 35 alternatives, 34 alternatives need further inspection to conclusively confirm whether they fall into Class 2 and 146 alternatives belong to other classes. The classification of all alternatives is effectively achieved. The effectiveness and feasibility of our approach is therefore validated.

6 Conclusions

We have designed a TWD method on MS-SVNDS in this paper. Considering the inadequacy of existing rough set models in describing uncertainty, we employ the theory of SVNSs, MSDSs and DS so as to construct a rough set model of MS-SVNDS. To address the knowledge acquisition issues involved in MS-SVNDS, we present the corresponding OPS and OPR approaches. Based on an optimal scale reduct of MS-SVNDS, we provide a novel TWD method to implement the classification. The experimental analysis is finally carried out to verify our method. The experimental results show that our method is reliable and effective.

The following list includes this paper's significant contributions and advantages:

  1. (1)

    A rough set model of MS-SVNDS is established. Compared with other existing models, our model has significantly improved the processing ability of inaccurate and uncertain information. At the same time, the model is closer to the real life needs and data collection way.

  2. (2)

    The OPS and OPR methods for MS-SVNDS are proposed. Through the above methods, we can obtain the simplified effective knowledge in MS-SVNDS according to the actual needs. And the classification performance based on the knowledge obtained by our method is better.

  3. (3)

    A TWD method on MS-SVNDS is given. The method differs from other methods in that it is based on simplified effective knowledge and the knowledge is in the form of SVNNs. More reliable classification and decision results can be obtained by using our method.

For our method, there exists two main limitations:

  1. (1)

    The introduction of SVNSs allows for a more detailed description of the data, but also leads to a much higher volume of data in pre-processing.

  2. (2)

    Our proposed TWD method is based on an obtained optimal scale reduct of MS-SVNDS. Although it can effectively implement classification, the operation is complicated.

Some of our considerations for future research topics are as follows:

  1. (1)

    We will further explore the integration of SVNSs and MSDSs.

  2. (2)

    We will further explore how to apply TWD method to the knowledge acquisition in MSDSs.