Abstract
The existing algorithms for data processing under probabilistic uncertainty often require too much computation time. Sometimes, we can speed up the corresponding computations if we take into account the fact that in many real-life situations, uncertainty can be naturally described as a combination of several components, components which are described by different granules. In such situations, to process this uncertainty, it is often beneficial to take this granularity into account by processing these granules separately and then combining the results. In this paper, we show that granular computing can help even in situations when there is no such natural decomposition into granules, namely we can often speed up processing of uncertainty if we first (artificially) decompose the original uncertainty into appropriate granules.
Similar content being viewed by others
References
Cormen TH, Leiserson CE, Rivest RL, Stein C (2009) Introduction to algorithms. MIT Press, Cambridge
Kovalerchuk B, Kreinovich V (2017) Concepts of solutions of uncertain equations with intervals, probabilities and fuzzy sets for applied tasks. Granul Comput 2(3):121–130
Kreinovich V (2016) Solving equations (and systems of equations) under uncertainty: how different practical problems lead to different mathematical and computational formulations. Granul Comput 1(3):171–179
Pedrycz W, Chen SM (2011) Granular computing and intelligent systems: design with information granules of high order and high type. Springer, Heidelberg
Pedrycz W, Chen SM (2015a) Information granularity, big data, and computational intelligence. Springer, Heidelberg
Pedrycz W, Chen SM (2015b) Granular computing and decision-making: interactive and iterative approaches. Springer, Heidelberg
Piegat A, Landowski M (2018) Solving different practical granular problems under the same system of equations. Granul Comput 3(1):39–48
Rabinovich SG (2005) Measurement errors and uncertainty: theory and practice. Springer, Berlin
Sheskin DJ (2011) Handbook of parametric and nonparametric statistical procedures. CRC, Boca Raton
Stylios CD, Pownuk A, Kreinovich V (2015) Sometimes, it is beneficial to process different types of uncertainty separately. In: Proceedings of the annual conference of the north American fuzzy information processing society NAFIPS’2015 and 5th world conference on soft computing, Redmond, Washington, August 17–19, 2015
Acknowledgements
The authors are thankful to Witold Pedrycz and Shyi-Ming Chen for their encouragement, and to the anonymous referees for valuable suggestions.
Funding
This article was funded by National Science Foundation (Grant nos. HRD-1242122, DUE-0926721) and Prudential (Grant no. and Prudential Actuarial Science Academy and Pipeline Initiative).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
On behalf of all authors, Vladik Kreinovich states that there is no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
This work was supported in part by the National Science Foundation Grants HRD-0734825 and HRD-1242122 (Cyber-ShARE Center of Excellence) and DUE-0926721, and by an award “UTEP and Prudential Actuarial Science Academy and Pipeline Initiative” from Prudential Foundation.
Appendix: Non-uniform distribution of \(\alpha _j\) is better
Appendix: Non-uniform distribution of \(\alpha _j\) is better
1.1 Idea
If we select two values \(\alpha _j\) too close to each other, there will be too much correlation between them, so adding the function corresponding to the second value does not add much information to what we know from a function corresponding to the first value.
We are approximating a general function (logarithm of a characteristic function) as a linear combination of functions \(|t|^{\alpha _j}\). If two values \(\alpha _j\) and \(\alpha _{j+1}\) are close, then the function \(|t|^{\alpha _{j+1}}\) can be well approximated by a term linear in \(|t|^{\alpha _j}\); thus, the term proportional to \(|t|^{\alpha _{j+1}}\) is not needed.
It, therefore, makes sense to select the values \(\alpha _j\) in such as way that for each j, the part of \(|t|^{\alpha _{j+1}}\) that cannot be approximated by terms proportional to \(|t|^{\alpha _j}\) should be the largest possible.
1.2 Let us reformulate this idea in precise terms
For every two functions f(t) and g(t), the part of g(t) which cannot be represented by terms \(a\cdot f(t)\) (proportional to f(t)) can be described as follows. It is reasonable to describe the difference between the two functions f(t) and g(t) by the least squares (\(L^2\)) metric \(\int (f(t)-g(t))^2\,{\text {d}}t\). In these terms, the value of a function itself itself can be described as its distance from 0, i.e., as \(\int (f(t))^2\,{\text {d}}t\).
When we approximate a function g(t) by a term \(a\cdot f(t)\), then the remainder \(g(t)-a\cdot f(t)\) has the value
The best approximation occurs when this value is the smallest, i.e., when it is equal to \(\min \nolimits _a \int (g(t)-a\cdot f(t))^2\,{\text {d}}t\). Out of the original value \(\int (g(t))^2\,{\text {d}}t\), we have unrepresented the part equal to \(\min \nolimits _a \int (g(t)-a\cdot f(t))^2\,{\text {d}}t\). Thus, the relative size of what cannot be represented by terms \(a\cdot f(t)\) can be defined as a ratio
1.3 Let us simplify the resulting expression
This expression can be simplified if we find the explicit expression for a for which the value \(\int (g(t)-a\cdot f(t))^2\,{\text {d}}t\) is the smallest possible. Differentiating the minimized expression with respect to a and equating the derivative to 0, we conclude that
i.e., that
and
For this a, the value \(\int (g(t)-a\cdot f(t))^2\,{\text {d}}t\) takes the form
Substituting the above expression for a into this formula, we conclude that
i.e., that
Thus, the desired ratio takes the form
Thus, we arrive at the following optimization problem.
1.4 Resulting optimization problem
To make sure that the above remainders are as large as possible, it makes sense to find the values \(\alpha ^{\mathrm{opt}}_1<\cdots <\alpha ^{\mathrm{opt}}_k\) that maximize the smallest of the remainders between the functions \(f(t)=|t|^{\alpha _j}\) and \(g(t)=|t|^{\alpha _{j+1}}\):
1.5 Solving the optimization problem
Let us consider an interval \([-T,T]\) for some T. Since the function is symmetric, it is sufficient to consider the values from [0, T].
For \(f(t)=t^\alpha \) and \(g(t)=t^\beta \), the integral in the numerator of the ratio is equal to
Similarly, the integrals in the denominator take the form
and
so
One can see that the powers of T cancel each other, and we get
or, equivalently, if we denote \(r{\mathop {=}\limits ^\mathrm{def}}\displaystyle \frac{\beta +0.5}{\alpha +0.5}\), we get
The derivative of the function R(r) is equal to
So this derivative is positive for all \(r>1\). Thus, the function R(r) is monotonically increasing, and looking for the values \(\alpha ^\mathrm{opt}_j\) for which \(\min \limits _j R(|t|^{\alpha _j},|t|^{\alpha _{j+1}})\) is the largest if equivalent to looking for the values \(\alpha ^{\mathrm{opt}}_j\) for which the smallest \(\min \limits _j \displaystyle \frac{\alpha _{j+1}+0.5}{\alpha _j+0.5}\) of the ratios \(r=\displaystyle \frac{\alpha _{j+1}+0.5}{\alpha _j+0.5}\) attains the largest possible value:
One can check that this happens when
Indeed, in this case, \(\min \nolimits _j \frac{\alpha _{j+1}+0.5}{\alpha _j+0.5}=\left( \frac{5}{3}\right) ^{1/(k-1)}.\) We cannot have it larger: if we had \(\min \nolimits _j \frac{\alpha _{j+1}+0.5}{\alpha _j+0.5}>\left( \frac{5}{3}\right) ^{k-1}\), then we would have \( \displaystyle \frac{\alpha _{j+1}+0.5}{\alpha _j+0.5}>\left( \displaystyle \frac{5}{3}\right) ^{k-1}\) for all j. Here
The first factor \(\alpha _1+0.5\) is \(\ge 1.5\), each of the other \(k-1\) terms is greater than \((\frac{5}{3})^{1/(k-1)}\), so for their product, we get
while we assumed that all the values \(\alpha _j\) are from the interval [1, 2], and so, we should have \(\alpha _k+0.5\le 2.5\).
1.6 Resulting optimal values of \(\alpha _j\)
Thus, the optimal way is not to take the values uniformly distributed in the interval [1, 2], but rather take the values
for which the logarithms \(\ln (\alpha ^\mathrm{opt}_j+0.5)= \frac{j-1}{k-1}\cdot \ln (\frac{5}{3})\) are uniformly distributed.
Comment It is worth mentioning that there is intriguing connection between these values \(\alpha _j\) and music: for example, the twelve notes on a usual Western octave correspond to the following frequencies:
-
the first note corresponds to the frequency \(f_1\),
-
the second note corresponds to the frequency
$$\begin{aligned} f_2=f_1\cdot 2^{1/12}, \end{aligned}$$ -
the third note correspond to the frequency
$$\begin{aligned} f_3=f_1\cdot 2^{2/12}, \end{aligned}$$ -
...,
-
the last note corresponds to the frequency
$$\begin{aligned} f_{12}=f_1\cdot 2^{11/12}, \end{aligned}$$and
-
the first note of the next octave corresponds to the frequency \(f_{13}=f_1\cdot 2\).
For these frequencies, the logarithms \(\ln (f_j)\) are uniformly distributed.
Similar formulas exist for five-note and other octaves typical for some Oriental musical traditions.
Rights and permissions
About this article
Cite this article
Pownuk, A., Kreinovich, V. Granular approach to data processing under probabilistic uncertainty. Granul. Comput. 6, 489–505 (2021). https://doi.org/10.1007/s41066-019-00210-5
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s41066-019-00210-5