Introduction

The field of information theory quantifies the information of an event by considering the concept of the extent of surprise that underlies an event. The events with low probability are generally had more surprising packages and information than the events with high probability because high-probability events are common in nature. The method used to calculate the amount of information that a random variable may possess is named Shannon information [1] and later on called entropy [2]. Clausius initially introduced entropy in the late nineteenth century by studying the dish gas disorder problem and quantifying the disorder produced by the gas through a measure named entropy [3]. The first statistical entropy measure proposed by Shannon [4] is defined as

$$ H\left( x \right)\; = \; - \int {f\left( x \right)} \;\log \left( {f\left( x \right)} \right)dx, $$
(1)

where \(f\left(x\right)\) is a pdf of a continuous probability distribution for the continuous random variable \(X\). Later on, other forms of entropy have been proposed and available in the literature are Shannon entropy [5], maximum entropy [6], Tsallis entropy [7], Topology entropy [8], Graph entropy [9], minimum entropy [10], approximate entropy [11], spectral entropy [12], sample entropy [13], fuzzy entropy [14], and cross-entropy measures [15]. Rényi entropy for order \(\alpha \) where \(\alpha \ge 0\) proposed by [16] is defined as

$$ H_{\alpha } \left( X \right)\; = \;\frac{1}{1 - \alpha }\;\log_{2} \sum {P_{i}^{\alpha } } \;{\text{for}}\;{\text{discrete}}\;{\text{distribusions,}} $$
(2)
$$ H_{\alpha } \left( X \right)\; = \;\frac{1}{1 - \alpha }\;\log_{2} \int {f\left( x \right)}^{\alpha } \;dx\;{\text{for}}\;{\text{continuous}}\;{\text{distribusions,}} $$
(3)

where \({P}_{i} (i=\mathrm{1,2},\dots ,n)\) are the probabilities for \(n\) possible outcomes of the r.v \(X\) and log2 is a logarithm to the base 2 operator.

The methods cited above for entropy measures consider only the deterministic outcome of an experiment or certain values and lacks in dealing with the interval form data. The interval form data arise from the experiment whose outcomes are not sure, and the exact outcome of the experiment is not certain. In such situations, the results of the experiment classified in interval form with some indeterminacy or vagueness. Smarandache [17] introduced the concept of fuzzy logic and neutrosophic statistics by incorporating vague, uncertain, unclear, or indeterminate states of data into mathematics and statistics theories.

It becomes challenging for the researchers to determine the exact sample size of a study when the responses for a survey are incomplete and confused for its inclusion. It is tough in real life to record the data precisely in crisp numbers, e.g., if the temperature of a particular area is to be measured, it is nearly impossible to estimate it precisely due to rapid fluctuations. Similarly, it is hard to obtain the water level of any river due to the same reason. Fuzzy techniques are great methods to deal with such types of situations. These techniques allow the statisticians to say that the particular room temperature is “around 29 °C” and the value 29 is considered a fuzzy number [18]. According to [19], the truthfulness or falseness of observations is just a matter of implementation of the degree of fuzzy logic, e.g., in the variables like tall, expensive, cold, low, etc. Zadeh introduced the theory of fuzzy sets in 1965 [19].

$$ T\left( p \right)\; = \;F\left( v \right), $$
(4)

where \(V\) is the variable that takes values “\(v\)” from the universal set \(V\), and \(F\) is the fuzzy set on \(V\) that represents vague predicate. If \(v\) is any particular value of \(V\), it is said to belong to the fuzzy group with membership grade \(F(v)\), the degree of truth, and \(T(p)\) of the proposition is used in (1).

Fuzzy sets played a central role in the investigation of non-deterministic techniques. In other words, fuzzy logic can help deal with imprecise environments. The concept of fuzzy logic in entropy measure was introduced by [14] and proposed a combination of fuzzy logic and entropy measure in characterizing the EMG (Electromyography) signals. This measure was a negative logarithm of conditional probability. [20] introduce the concept of neutrosophy and neutrosophic environment. The decision about the membership of interval numbers is proposed by Atanassove [21] under fuzzy environments through the interval-valued intuitionistic fuzzy set (IVIFS). Wei et al. [22] presented the concept of entropy measure for IVIFS in pattern identification and [23, 24] generalized the entropy measures for the neutrosophic set environment. Deli [25] developed a method for the linear optimization of the single-valued neutrosophic set and discussed its sensitivity. [26] presented a comparison of the attribute value and attributed weight single-valued trapezoidal neutrosophic numbers for the multiple attribute decision-making problems. Readers may find further details on neutrosophic set environment and entropy measures from [27,28,29,30,31,32,33,34,35]. However, instead of choosing a fuzzy number, we approach a more specific technique where both the numbers, in which lies the uncertainty, could be used. This approach is named neutrosophic statistics.

Neutrosophic statistics is a generalization of classical statistics because the methods of classical statistics are limited to only exact numbers and are helpless for indeterminate numbers or parameters [36]. [37] mentioned that fuzzy logic is a particular case of neutrosophic logic. Neutrosophic statistics is an extension to classical statistics known for its capacity to deal with values, i.e., values in which there is confusion, instead of real crisp whole numbers. In this type of statistics, we replace the indeterminate parameters and values (uncertain, unsure). Usually, it is represented by a capital “N” in the subscript of the parameter or the statistic such as \({x}_{N}\). Neutrosophic statistics deal with sets of values instead of crisp numbers. It replaces the parameters that are indeterminate with sets. Any number \(x\) replaced by a set-id denoted here as \({x}_{N}\) i.e. neutrosophic \(x\) or uncertain \(x\). This \({x}_{N}\) is mostly the interval including \(x\), generally, it is a set that approximates \(x\), see [38]. More information on neutrosophic Weibull distribution can be seen in [39] and [40].

The presence of inaccurate data, mainly in lifetime problems, is familiar, and the classical statistics are helpless in providing accurate estimates. These motivate the researchers to model the vague situation more precisely and present some methodologies capable of estimating the imprecise parameters. These sorts of situations could be dealt with neutrosophic sets where there is undoubtedly high and low value and uncertainty. Here in this study, neutrosophic logic and entropy are combined to provide a novel measure named neutrosophic entropy for an important lifetime distribution. Weibull distribution since the applications of neutrosophic theory with a combination of entropy measures has not been studied much in the past in statistical distribution theory. The developed methodologies will be beneficial in understanding the concept of entropy measures for the interval-valued data, mainly in the case of Weibull distribution.

Neutrosophic entropy of Weibull distribution

A neutrosophic set is a generalization of the classical fuzzy set. This set has the feature of fuzziness (truth, falsity, and indeterminacy) and imperfect information. Zadeh [41] used the entropy term to measure the fuzziness of a set. Details about entropy measures for fuzzy sets can be seen [2, 31, 42,43,44,45]. In this section, the neutrosophic entropy for a two-parameter Weibull distribution is derived. The distribution of a random variable X is known as the Weibull distribution with the shape parameter \(\alpha \) and scale parameter \(\beta \) with cumulative distribution function as [46]:

$$ F\left( x \right)\; = \;1 - {\text{e}}^{{ - \alpha x^{\beta } }} $$
(5)

and probability density function (pdf) as

$$ f\left( {x,\;\alpha ,\;\beta } \right)\; = \;\alpha \beta x^{\beta \; - \;1} {\text{e}}^{{ - \;\alpha x^{\beta } }} ,\;\;\;x\; > 0\;. $$
(6)

A neutrosophic form of the Weibull distribution with parameters [\({\alpha }_{L,}{\alpha }_{U}\)] and [\({{\beta }_{L},\beta }_{U}\)] is

$$ f\left( {x,\;\alpha_{{\left( {L,\;U} \right)}} ,\;\beta_{{\left( {L,\;U} \right)}} } \right) = \;\alpha_{{\left( {L,\;U} \right)}} \;\beta_{{\left( {L,\;U} \right)}} \;x^{{\beta_{{\left( {L,\;U} \right)}} \; - \;1}} {\text{e}}^{{ - \alpha_{{\left( {L,\;U} \right)}} x^{{\beta_{{\left( {L,\;U} \right)}} }} }} ,\;x > \;0. $$
(7)

Shannon entropy of Weibull distribution

In this section, Shannon entropy for Weibull distribution is proposed under the logic of neutrosophy. The Shannon entropy of Weibull distribution can be obtained by substituting (6) in (1), i.e.

$$ H\left( x \right)\; = \; - \log \left( {\alpha \beta } \right)\; + \;\frac{{\left( {\beta \; - 1} \right)}}{\beta }\left[ {\gamma \; + \;\log \left( \alpha \right)} \right]\; + 1, $$
(8)

where \(\gamma \) is the Euler’s constant and\(\gamma \;\;\int_{0}^{\infty } {{\text{e}}^{ - x} \ln x\;dx} .\)

Equation (8) shows the Shannon entropy of the Weibull distribution. This statistic is ideal for cases where we have crisp numbers and values for parameters. However, in real life, we scarcely come across crisp values, i.e., there is some uncertainty, and we are unable to decide among a pair of values such as the temperature does not have an exact value. It lies between the highest recorded temperature and the lowest at that time. The same is the case with the stock exchange records and numerous other cases in the real world. Neutrosophy as an extension of classical statistics could be helpful in situations having doubt or uncertainty among two values. To deal with such problems, we proposed a new entropy measure with the combination of neutrosophic statistics and named it neutrosophic entropy.

Using this Eq. (8), the neutrosophic Shannon entropy of Weibull distribution is

$$ H_{N} \left( x \right)\; = \;\left( { - \log \left( {\alpha_{L} \beta_{L} } \right) + \;\frac{{\left( {\beta_{L} \; - 1} \right)}}{{\beta_{L} }}\left[ {\gamma \; + \;\log \left( {\alpha_{L} } \right)} \right]\; + \;1,\; - \log \left( {\alpha_{U} \beta_{U} } \right) + \;\frac{{\left( {\beta_{U} \; - 1} \right)}}{{\beta_{U} }}\left[ {\gamma \; + \;\log \left( {\alpha_{U} } \right)} \right]\; + \;1} \right)\;. $$
(9)

Here, \({{\alpha }}_{L}\) is the estimate of the parameter \({\alpha }\) obtained from the lower bound for which the uncertainty lies in the sample; and \({\alpha }_{U}\) is the estimate of the parameter α of the Weibull distribution is obtained from the upper bound, among which lies the uncertainty in the sample. Similarly \({\beta }_{L}\) is the estimate of the parameter \(\beta \) of the Weibull distribution, estimated from the lower bounds in the sample also \({\beta }_{U}\) is the estimate of the parameter \(\beta \) for the Weibull distribution, estimated from the lower bounds in the sample. Where \({{\alpha }}_{L}{\upbeta }_{L}\) be the lower and \({{\alpha }}_{U}{\upbeta }_{U}\) be the upper-level values of the parameters. In circumstances where there appears doubt among more than one value (two values here), then the lower among them are considered as “L” whereas the more considerable value is “U”.

Using Eq. (5), we can obtain two results, one for the lower and the other for the upper bound, and we can state that the entropy in the particular case lies between these results.

Rényi entropy of Weibull distribution

The Rényi entropy of Weibull distribution is obtained by substituting (6) in the (3), i.e.

$$ H_{2} \left( X \right)\; = \frac{1}{1\; - \;2}\;\log_{2} \;\int_{1}^{\infty } {\left( {\alpha \beta x^{\beta \; - \;1} \;{\text{e}}^{{ - \alpha x^{\beta } }} } \right)}^{2} dx, $$
(10)
$$ H_{2} \left( X \right)\; = \log_{2} \;\left( {\left( {2\alpha } \right)^{\beta \; - \;1} \left| \!{\overline {\, {\left( \beta \right)} \,}} \right. } \right). $$
(11)

Equation (11) shows the Rényi entropy of the Weibull distribution. Entropy obtained from Eq. (11) is considered ideal when there is the availability of exact estimates of the parameters of Weibull distribution. As already discussed, it is infrequent in real life to obtain precise measures and estimates of parameters. There lies confusion or doubt, or the value cannot be precisely measured and calculated. Here is a need for a more general measure to calculate the entropy which is proposed here.

Neutrosophic Rényi entropy of Weibull distribution

The neutrosophic Rényi Entropy for Weibull distribution using Eq. (11) is

$$ H_{2N} \left( X \right)\; = \;\log_{2} \left( {\left( {2\alpha_{L} } \right)^{{ - \beta_{L} }} \beta_{L} } \right)\left| \!{\overline {\, {\left( {\beta_{L} } \right)} \,}} \right. ,\;\log_{2} \left( {\left( {2\alpha_{U} } \right)^{{ - \beta_{U} }} \left| \!{\overline {\, {\left( {\beta_{U} } \right)} \,}} \right. } \right), $$
(12)

where \({{\alpha }}_{L}{\upbeta }_{L}\) be the lower and \({{\alpha }}_{U}{\upbeta }_{U}\) be the upper-level estimates of the parameters. In circumstances where there appears doubt among more than one value (two values here), then the lower among them are considered as “L” whereas the larger value is “U”.

Using Eq. (5), we can obtain two results, one for the lower and the other for the upper bound, and we can state that the entropy in the particular case lies between these results.

Simulation study

In this section, a simulation study is performed to assess the effectiveness and efficiency of the proposed measure. The steps used for the simulation study are

  1. 1.

    The theoretical neutrosophic Shannon and Rényi entropy for some chosen intervals of the shape parameter \(\alpha \) and scale parameter \(\beta \) in the interval form were computed.

  2. 2.

    Random numbers x1 and x2 were generated from the Weibull distribution of sample size 10,000.

  3. 3.

    Compute the estimates of the shape parameter \(\alpha \) and scale parameter \(\beta \) for the lower and upper bound from the obtained samples.

  4. 4.

    Now HN(x) can be attained using Eq. (9) and H2N(x) can be achieved using Eq. (12).

  5. 5.

    We repeat the steps mentioned above for 1000 random samples and different parameters to check the efficacy of the novel measure.

  6. 6.

    The empirical neutrosophic Shannon and Rényi entropy for Weibull distribution were computed by taking the average of the obtained matrix of the obtained entropies in the interval form.

  7. 7.

    The theoretical and empirical neutrosophic entropies were compared and the obtained results were plotted.

The empirical results obtained from simulations for the neutrosophic Shannon entropy of the Weibull distribution are shown in Table 1 for different values of α and β. First, \(\alpha \) is fixed and variations due to \(\beta \) were observed. It can be observed from Table 1 that there are some patterns and relations among the values of estimates of parameters and proposed entropy measure. It can be seen that on increasing the values of parameters, the neutrosophic Shannon entropy slightly increases. Second, the parameter \(\beta \) is fixed and variations due to \(\alpha \) are observed. It is observed that there is a slight increase in the lower bound of neutrosophic Shannon entropy on increasing the values of parameters, whereas the upper bound remains the same.

Table 1 Neutrosophic Shannon entropy for different values of α and β

Similarly, Table 2 shows the empirical results, obtained from simulation, of neutrosophic Rényi entropy of Weibull distribution, for different values of α and β. It can be seen from Table 2 that on increasing the value of parameter \(\beta \) the neutrosophic entropy increases. It can also be stated that it approaches zero. It can also be seen from Table 2 that on increasing the value of parameter \(\alpha \) there is very little increase in the upper bound of neutrosophic entropy while the lower bound remains constant.

Table 2 Neutrosophic Rényi entropy for different values of α and β

Figure 1 shows the neutrosophic Shannon and Rényi entropy for Weibull distribution for \(\alpha \) = (76.25, 97.2) and \(\beta \) = (9.95, 12.05). The red line represents the lower bound of the obtained interval of neutrosophic entropy while the blue line represents the lower bound of the obtained interval of neutrosophic entropy of the Weibull distribution. In this situation, there exists indeterminacy among an interval of values of parameter, i.e. \(\alpha \) = (76.25, 97.2) and \(\beta \) = (9.95, 12.05), so the neutrosophic entropy is suitable here. Figure 1 shows that the value (true value) of entropy lies in the obtained interval.

Fig. 1
figure 1

The neutrosophic Shannon and Rényi entropy for Weibull distribution for \(\alpha \) = (76.25, 97.2) and \(\beta \) = (9.95, 12.05)

Applications on real-life data sets

Example 1:

We have taken the data of the past 5 years, i.e., 2016–2020 temperatures (Average Low and Average High) of Lahore, Punjab, Pakistan. The data include a forecast of the remaining months of 2020, i.e. (September–December). The data were collected from World Weather Online [47]. Figure 2 shows the graph of the highest recorded temperature in Lahore with an orange line. The average temperature recorded is represented by the blue line, while the black line represents the lowest recorded temperature.

Fig. 2
figure 2

Temperature of Lahore (oF) for 2016–2020

Here temperatures for five years are observed, i.e., from the year 2016–2020. Using these observations the estimates of the neutrosophic Weibull parameters were obtained. The analysis was done in Minitab (Ver. 18), and the parameters of neutrosophic Weibull distribution (Eq. 7) were estimated. These estimates are used to find the neutrosophic Shannon entropy of the Weibull distribution (Eq. 9). Graph (1) shows these estimates.

From Table 3 in the Appendix and Fig. 3, the values of \({{\alpha }}_{L}{{ and \beta }}_{L}\) be the lower and \({{\alpha }}_{U} and {\upbeta }_{U}\) be the upper-level estimates of the interval parameters \({\alpha and \beta }\), (\({{\alpha }}_{L}{{\alpha }}_{U}\)) = (76.35, 97.22) also (\({\upbeta }_{L}{\upbeta }_{U}\)) = (5.770, 8.544) and N = 60. The result obtained for the neutrosophic Shannon entropy of Weibull distribution using Eq. 9 is \({H}_{N}\left(x\right)=(0.3897, 0.3451)\) and the result obtained for the neutrosophic Rényi entropy of Weibull distribution using Eq. 12 is, i.e., H2N(x)\(=(-0.005,-0.0039)\). The study consists of average low and average high temperatures of Lahore, Punjab, Pakistan for the past five years (including 2020). The data also incorporate the weather forecast for the remaining months of 2020. These data were collected from World Weather Online. To analyze the data, Minitab (ver. 18) software was used. The results showed that the entropy value is not so large, i.e., close to 1 and quite far from 0. It is said that entropy value lies mostly between 0 and 1. Zero means no uncertainty or error, whereas one states maximum uncertainty and unpredictability; in other words, we can say a very high level of disorder or low purity level. From this example, we get the information that the system indulged in estimating the average monthly temperatures of a particular city, country, or area (Lahore, Punjab, Pakistan here) is accurate, and it has a significantly less amount of disorder. The neutrosophic Shannon entropy obtained here is close to zero, i.e. (0.3897, 0.3451), which indicates less uncertainty in the data taken and the system used to collect this kind of data efficiently. It can also be said that the weather condition of Lahore is not that uncertain as the entropy value is close to zero, and the value of H(x) close to zero indicates less amount of uncertainty. The neutrosophic Rényi entropy obtained here is close to zero, i.e. (− 0.005, − 0.0039), indicating less uncertainty in the data taken and the system used to collect this kind of data is efficient. It can also be said that the weather condition of Lahore is not that uncertain as the entropy value is close to zero, and the value of H(x) close to zero indicates less amount of uncertainty.

Fig. 3
figure 3

Histogram of temperatures of Lahore (2016–2020)

Example 2:

Lieblein and Zelen [48] give the results of tests of the endurance of nearly 5000 deep-groove ball bearings. The graphical estimates of cover all lots tested appear to have an average value of about 1.6. Consider the following sample given the results of the tests, in millions of revolutions, of 23 ball bearings were 17.88, 28.92, 33.00, 41.52, 42.12, 45.60, 48.48, 51.84, 51.96, 54.12, 55.56, 67.80, 68.64, 68.64, 68.88, 84.12, 93.12, 98.64, 105.12, 105.84, 127.92, 128.04, 173.40. The neutrosophic form data are provided in the Appendix.

The maximum likelihood estimate of \(\beta \) is 2.102, the estimate of \({\alpha }\) from the equation is 81.99. The upper-level estimates were randomly generated using Weibull distribution with parameters alpha = 85.99 and beta = 2.259, keeping n = 23.

From Table 4 in the Appendix, the values of \({{\alpha }}_{L}{{ and \beta }}_{L}\) be the lower and αU\( and {\upbeta }_{U}\) be the upper-level estimates of the parameters α\({ and \beta }\), (\({{\alpha }}_{L}{{\alpha }}_{U}\)) = (81.99, 85.99) also (\({\upbeta }_{L}{\upbeta }_{U}\)) = (2.102 ,2.259) and n = 23. The result obtained is the neutrosophic Shannon entropy and Rényi entropy of Weibull distribution, i.e. \({H}_{N}\left(x\right)=( -0.2505884 1.3711825)\) and H2N(x)\(=(3.991186, 17.589184)\).

Figure 4 shows the neutrosophic Shannon and Rényi entropy for Weibull distribution for \(\alpha \) = (81.99, 85.99) and \(\beta \) = (2.102, 2.259). The red line represents the lower bound of the obtained interval of neutrosophic entropy. In contrast, the blue line represents the lower bound of the obtained interval of neutrosophic entropy of the Weibull distribution. In this situation there exists indeterminacy among an interval of values of parameter i.e. \(\alpha \) = (81.99, 85.99) and \(\beta \) = (2.102, 2.259) so the neutrosophic entropy is suitable here. The Fig. 3 shows that the value (true value) of entropy lies in the obtained interval. The Shannon entropy and Renyi entropy estimates under classical statistics for the parameters α = 83 and β = 2.15 are \(H(x)=-2.5\), H2(x) = 4.07, respectively, that are quite far from the estimates under the proposed forms of Shannon and Renyi entropies. These indicate that our proposed methodologies are better than the classical methods under a neutrosophic environment.

Fig. 4
figure 4

The neutrosophic Shannon and Rényi entropy for Weibull distribution for \(\alpha \) = (81.99,85.99) and \(\beta \) = (2.102, 2.259)

Example 3:

The data in Table 5 are the times to failure measurement of 25 lamp bulbs on accelerated test. The data were taken from Wadsworth [49] (shape parameter 1.985, and scale = 0.976). The upper-level estimate was estimated with shape parameter = 2.220 and scale parameter = 0.999. Also, \({H}_{N}(x)=( 0.312, 1.023)\) and H2N(x)\(=\left(0.965, 1.989\right).\)

Figure 5 shows the neutrosophic Shannon and Rényi entropy for Weibull distribution for \(\alpha \) = (1.985, 2.220) and \(\beta \) = (0.976, 0.999). The red line represents the lower bound of the obtained interval of neutrosophic entropy, while the blue line represents the lower bound of the obtained interval of neutrosophic entropy of the Weibull distribution. In this situation there exists indeterminacy among an interval of values of parameter i.e. \(\alpha \) = (1.985, 2.220) and \(\beta \) = (0.976, 0.999). The Shannon entropy and Renyi entropy estimates under classical statistics for the parameters α = 1.996 and β = 0.986 are \(H(x)=1.01\), H2(x) = 0.98, respectively, are far from the estimates under the proposed forms of Shannon and Renyi entropies. These indicate that our proposed methodologies are better than the classical methods under a neutrosophic environment. Figure 3 shows that the value (true value) of entropy lies in the obtained interval.

Fig. 5
figure 5

The neutrosophic Shannon and Rényi entropy for Weibull distribution for \(\alpha \) = (1.985, 2.220) and \(\beta \) = (0.976, 0.999)

Concluding remarks

Entropy is an essential measure from information theory to determine the vagueness of a data set in an exact form. The measure is also necessary for the distribution theory as a measure of uncertainty. Several studies have been conducted to fix the problem in distribution theory were based on the numbers generated in an exact form and unable to solve the problems having interval-valued data. Although some entropy measures available in the literature deal with the interval-valued data but in sets form data and lacks in studying a probability distribution having interval-valued data. This research fills the gap by setting the entropy measures in distribution theory in the context of neutrosophy and interval-valued data. Important entropy measures, i.e., Shannon and Renyi, were modified for the Weibull distribution under the neutrosophic interval-valued logic. The proposed entropy methods were the generalization of the existing entropy methods under classical statistics. We showed both simulation and real data set applications that the proposed modified forms of the entropies measure for Weibull distribution produced efficient estimates in the presence of interval-valued data. As the Weibull distribution is a vital lifetime distribution in probability theory and has wider applications in engineering, quality control, medical sciences, etc., we can say that our proposed forms of entropy measures also have broad applications problems generated by the interval-valued data. The concept and applications of the proposed entropy measures can be extended for other neutrosophic probability distributions from the distribution theory. Moreover, other entropy measures available in the literature can be generalized under the neutrosophic environment for a more comprehensive application of entropy theory.