1 Introduction

Time series is a sequence of real-data, with each element in this sequence representing a value recorded at some time moment. With regard to processed data, time series is inherently associated with their large size, high dimensionality and a stream-like nature. Time series are ubiquitous in numerous domains. Recently, this data modality has attracted great attention in the data mining research community. In the plethora of currently available models of time series, the accuracy of the models has been a holy grail of the overall modeling. With the emergence of a more visible need for interpretable models that are easily comprehended by users and facilitate a more interactive style of modeling, arose an important need to develop models that are not only accurate but transparent (interpretable) Casillas et al. (2003) as well. Likewise, as a part of requirements, it is anticipated that the user should be at position to easily establish a sound tradeoff between accuracy and interpretability. Similarly, effective mechanisms supporting a realization of relevance feedback become an important part of the overall research agenda.

Current research in time series mining, time series characterization, and prediction comes with a visible diversity and richness of the conceptual and algorithmic pursuits. In spite of their diversity, there is some evident and striking similarity; all of these approaches dwell upon the use and processing being realized at a numeric level.

The pursuits of perception, analysis and interpretation of time series, as realized by humans, are realized at a certain, usually problem-related, level of detail. Instead of single numeric entries—successive numeric readings of time series, developed are conceptual entities over time and feature space —information granules using which one discovers meaningful and interpretable relationships forming an overall description of time series. The granularity of information is an important facet being imperative to any offering of well-supported mechanisms of comprehension of the underlying temporal phenomenon. In all these pursuits, information granules manifest along the two main dimensions (as noted above). The first one is concerned with time granularity. Time series is split into a collection of time windows —temporal granules. One looks at time series in temporal windows of months, seasons, years. Time series is also perceived and quantified in terms of information granules being formed over the space of amplitude of the successive samples of the sequence; one arrives at sound and easily interpretable descriptors such as low, medium, high amplitude and alike. One can also form information granules over the space of changes of the time series. Combined with the temporal facets, the composite information granules arise as triples of entities, say long duration, positive high amplitude, approximately zero changes, etc. Once information granules are involved, long sequences of numbers forming time series are presented as far shorter, compact sequences of information granules-conceptual entities. As noted above, those are easier to understand and process. Let us note that while temporal granules are quite common (and they are of the same length), forming and processing composite information granules built over variable length time intervals call for detailed investigations.

Let us emphasize that information granules are sought as entities composed of elements being drawn together on a basis of similarity, functional closeness or spatial neighborhood. The quality of such granulation (abstraction) of data is clearly related to the ability of this abstraction process to retain the essence of the original data (problem) while removing (hiding) all unnecessary details. With this regard, granularity of information (Bargiela and Pedrycz 2002, 2003, 2005, 2008, 2009; Apolloni etal. 2008; Srivastava et al. 1999; Ślȩzak 2009; Pedrycz and Song 2011; Qian et al. 2011) plays a pivotal role and becomes of paramount importance, both from the conceptual as well as algorithmic perspective to the realization of granular models of time series. Subsequently, processing realized at the level of information granules gives rise to the discipline of granular computing (Bargiela and Pedrycz 2003). In granular computing, we encounter a broad spectrum of formal approaches realized in terms of fuzzy sets (Aznarte et al. 2010; Chen and Chen 2011; Kasabov and Song 2003; Lee et al. 2006; Pedrycz and Gomide 2007), rough sets (Pawlak 1991, 1985; Pawlak and Skowron 2007a, b), shadowed sets (Pedrycz 1998, 1999), probabilistic sets (Hirota and Pedrycz 1984), and others. Along with the conceptual setups, we also encounter a great deal of interesting and relevant ideas supporting processing of information granules. For instance, we can refer to the algorithmic perspective embracing fuzzy clustering (Bezdek 1981), rough clustering, and clustering being regarded as fundamental development frameworks in which information granules are constructed.

Fig. 1
figure 1

TAIEX time series and the corresponding temporal windows

Fig. 2
figure 2

IBM stock price time series and the corresponding temporal windows

The key objective of this paper is to develop a general architecture of a system supporting human-centric analysis and interpretation of time series, both through their visualization as well as a linguistic description. In both cases, we carefully exploit concepts of information granules as being central to all faculties of description and interpretation of time series. In order to form an abstract, concise view at temporal data, the associated information granule is made justifiable, viz. strongly supported by experimental evidence, and semantically meaningful (of sufficient specificity, sound level of detail). We briefly discuss data sets used in Sect. 2. Section 3 recalls the underlying processing issues of time series. The granular framework of interpretation of time series is proposed in Sect. 4. A crux of a principle of justifiable granularity is outlined in Sect. 5. The principle gives rise to an information granulation algorithm, which takes a collection of temporal data and transforms them into a semantically sound information granule. It is shown how interval granules are built and how a collection of such nested granules are arranged into a fuzzy set. Section 6 is concerned with a formation of a collection of information granules. Composite information granules formed in the space of amplitude and changes of amplitude are studied. A suite of experimental studies comprising several real-world time series is discussed in Sect. 6. Conclusions are offered in Sect. 7.

A certain point has to be made with regard to the formalism of granular computing being used in this study. While most of the detailed investigations are carried out in the setting of fuzzy sets (and in this way one could have contemplated the usage of the term fuzzy time series as a more suitable wording), which become instrumental in all the pursuits here, offering a conceptual framework and facilitating a realization of the sound algorithmic platform. It is emphasized that the introduced layered approach is equally suitable to deal with a variety of formalisms of information granules as well as the proposed constructs (such as e.g., the principle of justifiable granularity supporting a formation of a variety of information granules) and hence the notion of granular time series becomes well justified.

2 Experimental data sets—a summary

In this study, three publicly available time series coming from the finance domain are used to visualize a way how the granular constructs are designed and emphasize their essential aspects. More specifically, we consider here:

(1) The daily value of Taiwan Stock Exchange Weighted Stock Index (TAIEX) time series from 4th January 2000 to 29th December 2000, see Fig. 1a (sources: http://finance.yahoo.com/q/hp?s=%5ETWII&a=00&b=1&c=2000&d=11&e=29&f=2000&g=d ).

(2) The daily value of IBM common stock closing prices from 29th June 1959 to 30th June 1960, see Fig. 2a (sources: http://datamarket.com/data/set/2321/ibm-common-stocklosing-prices-daily-29th-jun%e-1959-to-30th-june-1960-n255#!display=table&ds=2321).

(3) The daily value of NASDAQ Composite Index (Nasdaq, for short) time series from 3th January 2000 to 31th December 2012, see Fig. 3a (sources: http://finance.yahoo.com/q/hp?s=%5EIXIC&a=00&b=1&c=2000&d=11&e=31&f=2012&g=d ).

Fig. 3
figure 3

Nasdaq time series and the corresponding temporal windows

For each time series, the changes (first order differences) of the amplitude are smoothed with the use of the moving average filter (Brown 2004); the obtained results is presented in Figs. 1b, 2b and 3b using a dotted line. Besides, the temporal windows are also showed in Figs. 1,  2 and  3: the first two time series are split into 20 temporal windows of the same length and the third time series is split into 30 temporal windows of equal length. These values of the size of the windows are selected for illustrative purposes.

3 Time series-selected underlying processing issues

For the purpose of reduction of data and facilitating all mining algorithms, in most development schemes discussed is a segmentation step. Time series segmentation can be treated either as a preprocessing stage for a variety of ensuing data mining tasks or as an important stand-alone analysis process. An automatic partitioning of a time series into an optimal (or better to say, feasible) number of homogeneous temporal segments becomes an important problem. Quite often considered is a fixed-length segmentation process. Commonly encountered segmentation methods include the perceptually important points (PIP) (Fu et al. 2006; Jiang et al 2007), minimum message length (MML) (Oliver et al. (1998)) and minimum description length (MDL) segmentation (Fitzgibbon et al. 2002). A two-stage approach, in which one first uses a piecewise generalized likelihood ratio to carry out rough segmentation and then refines the results, has been proposed by Yager (2009). Keogh et al. (2001) adopted a piecewise linear representation method to segment time series. They focused on the problem of an-online segmentation of time series where a sliding window and bottom-up approach was proposed. Fuzzy clustering algorithms have showed a significant potential to address this category of problems. With this regard, Abonyi et al. (2005) developed a modified Gath-Geva algorithm to divide time-varying multivariate data into segments by using fuzzy sets to represent such temporal segments. Duncan and Bryant (1996) proposed dynamic programming to determine a total number of intervals within the data, the location of these intervals and the order of the model within each segment. Wang and Willett (2002) would be the segmentation problem as a tool for exploratory data analysis and data mining called the scale-sensitive gated experts (SSGE), which can partition a complex nonlinear regression surface into a set of simpler surfaces called “features”. Fu (2011) presented a recent survey of mining time series.

4 A granular framework of interpretation of time series: a layered approach to the interpretation of time series

As noted so far, the notion of information granularity plays a pivotal role in all interpretation and analysis pursuits of time series. Our investigations of the description of time series being cast in the setting of granular computing are presented in a top-down fashion. We start with an overall view of the conceptual framework by stressing its key functionalities and a layered architecture and then move on with a detailed discussion by elaborating on the supported algorithmic aspects.

As commonly encountered in the investigations on time series, a starting point is a collection of numeric data (samples recorded in successive time moments) time series \(\{x_1, x_2\dots x_N\}\). For the purpose of further analysis, we also consider the first-order dynamics of the time series by considering the sequences of differences (changes) observed there, namely \(\{\Delta x_2, \Delta x_3 \dots \Delta x_N\}\) where \(\Delta x_i = x_i-x_{i-1}\). Quite often a smoothed version of these differences is sought. The space in which time series are discussed comes as a Cartesian product of the space of amplitude \(X\) and change of amplitude, namely \(X\times \Delta X\).

Fig. 4
figure 4

Development of granular time series: visualization and linguistic interpretation. Emphasized is a layered hierarchical approach presented in this study

The bird’s-eye view of the overall architecture supporting processing realized at several layers which stresses the associated functionality is displayed in Fig. 4. Let us elaborate in detail on the successive layers at which consecutive phases of processing are positioned:

Formation of information granules The granules are formed over the Cartesian product of amplitude and change of amplitude, viz. the space \(X\times \Delta X\). Here for each time slice \(T_i\) (note that the time variable is also subject to granulation as well; in addition to the same length time intervals, their size could be eventually optimized), we form an interval information granule (Cartesian product) \(X_i \times \Delta X_i\). These information granules are constructed following the principle of justifiable granularity, as outlined in Sect. 5.1 and discussed further in Appendix A.

Visualization of information granules The results produced at the previous processing phase are visualized. In essence, for each time slice (segment) \(T_i\), one can visualize a collection of Cartesian products of the information granules in \(X\times \Delta X\) obtained in successive time slices, see Fig. 4.

Linguistic description of granular time series While the visualization of the granular time series could be quite appealing highlighting the main temporal tendencies observed in the time series, it is also worth arriving at the linguistic description of the time series, which is articulated in terms of some linguistic (granular) landmarks and ensuing levels of the best matching of these landmarks with the constructed information granules \(X_i \times \Delta X_i\). Having a collection of the landmarks \(A_i\), \(i=1,2,\ldots ,c\) where typically \(c\ll p\) (where “\(p\)” stands for the number of time slices of the time series), the linguistic description comes as a string of the landmarks (for which the best matching has been accomplished) along with the corresponding matching levels \(\mu _1,\mu _2,\ldots ,\mu _p\). For instance, the linguistic description of the time series can read as the following sequence of descriptors:

{(positive small,negative medium})(0.7)}

{(negative large,around zero)(0.9)}\(\dots \)

In the above characterization, each granular landmark comes with its own semantics (e.g., positive small, negative medium, etc). The triples linguistically describe the amplitude and its change (expressed in terms of linguistic terms), and the associated matching level. To arrive at the linguistic description of this nature, two associated tasks are to be completed, namely (a) a construction of meaningful (semantically sound) granular landmarks, and (b) invocation of a matching mechanism, which returns a degree of matching achieved. The first task calls for some mechanism of clustering of information granules while the second one is about utilizing one of the well-known matching measures encountered in fuzzy sets, say possibility or necessity measures.

Linguistic prediction models of time series The descriptions of the time series are useful vehicles to represent (describe) time series in a meaningful and easy to capture way. Per se, the descriptors are not models such as standard constructs reported in time series analysis. They, however, deliver all components, which could be put together to form granular predictive models. Denoting by \(A_1,A_2,\ldots ,A_c\) the linguistic landmarks developed at the previous phase of the overall scheme, a crux of the predictive model is to determine relationships present between the activation levels of the linguistic landmarks present for the current time granule \(T_k\) and those levels encountered in the next time granule \(T_{k+1}\). The underlying form of the predictive mapping can be schematically expressed in the following way,

$$\begin{aligned} \begin{array}{l} A_1(X_k),A_2(X_k),\ldots ,A_c(X_k)\rightarrow A_1(X_{k+1}),\\ A_2(X_{k+1}),\ldots ,A_c(X_{k+1}) \end{array} \end{aligned}$$
(1)

where \(A_i(X_k)\) stands for a level of activation (matching) observed between \(A_i\) and the current information granule \(X_k\). The operational form of the predictive model can be realized in the form of a fuzzy relational equation

$$\begin{aligned} A(X_{k+1})=A(X_k)\circ R \end{aligned}$$
(2)

with “\(\circ \)” being a certain composition operator used in fuzzy sets (say, max-min or max-t composition where “t” stands for a certain t-norm) completed over information granules. \(A_i(X_{k+1})\) is the activation levels of the linguistic landmarks \(A_i\) caused by the predicted information granule \(X_{k+1}\). Overall, the vector \(A(X_k)\) has “\(c\)” entries, namely \([A_1(X_k) A_2(X_k)\dots A_c(X_k)]\). The granular relation \(R\) of dimensionality \(c\times c\) captures the relationships between the corresponding entries of \(A(X_k)\) and \(A(X_{k+1})\). Note that as a result of using the model  2, \(X_{k+1}\) is not specified explicitly but rather through the degrees of activation of the linguistic landmarks. In other words, the predictive granular model returns a collection of quantified statements (the corresponding entries of \(A(X_{k+1}))\)

$$\begin{aligned} \begin{array}{ll} &{}\text {--predicted information granule is } A_1 \text { with degree of} \\ &{}\text {activation } \lambda _1(=A_1(X_{k+1}))\\ &{}\text {--predicted information granule is } A_2 \text { with degree of} \\ &{}\text {activation } \lambda _2(=A_2(X_{k+1}))\\ &{}\dots \\ &{}\text {--predicted information granule is } A_c \text { with degree of} \\ &{}\text {activation } \lambda _c(=\!\!A_c(X_{k+1}))\\ \end{array} \end{aligned}$$
(3)

which offer a certain intuitive view at the prediction outcome. Obviously, one can easily choose the dominant statement for which the highest level of matching has been reported, namely predicted information granule is \(A_i\) with degree of activation \(\lambda _i(\!=\!\!\!A_i(X_{k+1}))\) where \(\lambda _i=arg~max_{j=1,2,\ldots ,c}A_j(X_{k+1}))\).

Note that there might be also another relational dependency as to the predicted length of the next time granule Tk+1 (in case the sizes of temporal information granules vary across the description of the time series)

$$\begin{aligned} A_1(X_k),A_2(X_k),\ldots ,A_c(X_k),T_k \rightarrow T_{k+1} \end{aligned}$$
(4)

which can be regarded as a certain granular relational equation describing dependencies among time granules \(T_k\) and \(T_{k+1}\)

$$\begin{aligned} T_{k+1}=A(X_k)\circ T_k \circ G \end{aligned}$$
(5)

In the above expression, \(G\) stands for a fuzzy relation, which links the input and output variables while “\(\circ \)” denotes a sup-t composition operator. Alluding to expressions  (1)–(5), we can view these predictive models as granular models of order-2 as they are described over the space of information granules rather than numeric entries.

5 Construction of information granules

The processing activities supporting the realization of information granules outlined in the previous section are now discussed in detail by presenting the successive phases of this process. We start with a brief elaboration on the principle of justifiable granularity, which forms a conceptual and algorithmic setup for the formation of information granules in the presence of some experimental evidence (numeric data).

5.1 The principle of justifiable granularity

We are concerned with a development of a single information granule (interval) \(\Omega \) based on some numeric experimental evidence coming in a form of a collection of a one-dimensional (scalar) numeric data, \(D=\{x_1,x_2 \dots x_N\}\). The crux of the principle of justifiable granularity is to construct a meaningful information granule based on available experimental evidence (data) \(D\). The require that such a construct has to adhere to the two intuitively compelling requirements:

(a) Experimental evidence The numeric evidence accumulated within the bounds of \(\Omega \) has to be made as high as possible. By requesting this, we anticipate that the existence of the constructed information granule is well motivated (justified) as being reflective of the existing experimental data. The more data are included within the bounds of \(\Omega \), the better—in this way the set becomes more legitimate (justified).

(b) Semantics of the construct At the same time, the information granule should be made as specific as possible. This request implies that the resulting information granule comes with a well-defined semantics (meaning). In other words, we would like to have \(\Omega \) highly detailed, which makes the information granule semantically meaningful (sound). This implies that the smaller (more compact) the information granule (lower information granule) is, the better. This point of view is in agreement with our general perception of knowledge being articulated through constraints (information granules) specified in terms of statements such as “\(x\) is \(A\)”, “\(y\) is \(B\)”, etc. where \(A\) and \(B\) are constraints quantifying knowledge about the corresponding variables. Evidently, the piece of knowledge coming in the form “\(x\) is in \([1,3]\)” is more specific (semantically sound, more supportive of any further action, etc.) than another less detailed piece of knowledge where we know only that “\(x\) is in \([0,12]\)”.

The further discusses for the principle of justifiable granularity is detailed in Appendix A.

5.2 Formation of fuzzy sets out of a family of interval information granules

For the data contained in the time interval \(T_i\), we consider samples of time series \(\{x_1,x_2 \dots x_N\}\) and their differences \(\{\Delta x_1, \Delta x_2 \dots \Delta x_{N-1}\}\) out of which following the principle outlined in Sect. 5.1, we construct interval information granules \(X_i\) and \(\Delta X_i\) and their Cartesian product. As these Cartesian products are indexed by the values of \(\alpha \), as a results of combining a family of the Cartesian products we form a fuzzy set-represented as a family of \(\alpha \)-cuts \((X_i \times \Delta X_i)_\alpha \).

6 Clustering information granules—a formation of linguistic landmarks

Having a collection of information granules, they are clustered to form some meaningful representatives with well-expressed semantics. The fuzzy C-Means algorithm is one of viable alternatives to be used here. Schematically, we can illustrate the overall process as shown in Fig. 5.

Fig. 5
figure 5

From granular data to linguistic landmarks

Fig. 6
figure 6

Granular prototypes of TAIEX time series in the \(x-\Delta x\) feature space

Fig. 7
figure 7

Granular prototypes of IBM stock prices time series in the \(x-\Delta x\) feature space

The information granules, a collection of \(\alpha \)-cuts formed within the time intervals are clustered to build meaningful semantically sound descriptors of the time series. There are two aspects of the formation of information, which deserve attention: (a) feature space in which clustering takes place. Given the nature of objects (information granules) to be clustered, which are undoubtedly more sophisticated than plain numeric entities, the granular data are expressed by a finite number of their \(\alpha \)-cuts; for the pertinent details see Appendix B. The clustering is realized separately for each variable (namely amplitude and change of amplitude). This helps us order the obtained prototypes in a linear way and associate with them some semantics, say low, medium, high, etc. (b) The granular prototypes formed for the individual variables give rise to compound descriptors such as e.g., amplitude high and change of amplitude close to zero. This is accomplished by forming Cartesian products of the granular prototypes, say \(A_i\times B_j\) with \(A_i\) and \(B_j\) being the granular prototypes already constructed in the space of amplitude \(x\) and its change \(\Delta x\). In general having \(r_1\) and \(r_2\) prototypes in the corresponding spaces, forming all prototypes constructed over the individual variables, we end up with \(r_1 r_2\) Cartesian products, viz. compound information granules. Some of them could be spurious that is there is no experimental legitimacy behind them (a very few data associated with the cluster). Thus once the Cartesian product has been formed, its legitimacy vis-à-vis experimental data needs to be checked and quantified. The quantification is realized in the form of a weighted counting of the number of data “covered by granular prototype”, for details to see Appendix B.

Fig. 8
figure 8

Granular prototypes of Nasdaq time series in the \(x-\Delta x\) feature space

Table 1 TAIEX time series: linguistic terms associated with granular prototypes
Table 2 TAIEX time series: linguistic description of granular prototypes
Table 3 IBM stock prices time series: linguistic terms associated with granular prototypes
Table 4 IBM stock prices time series: linguistic description of granular prototypes
Table 5 Nasdaq time series: linguistic terms associated with granular prototypes
Table 6 Nasdaq time series: linguistic description of granular prototypes
Table 7 TAIEX time series: description in terms of granular prototypes (in brackets shown are the prototypes with the second highest level of matching)
Table 8 IBM stock prices time series: description n terms of granular prototypes (in brackets shown are the prototypes with the second highest level of matching)
Table 9 Nasdaq time series: description in terms of granular prototypes (in brackets shown are the prototypes with the second highest level of matching)

Proceeding with the collection of the experimental data, the resulting granular prototypes by taking Cartesian product are displayed respectively in Figs. 6a, 7a and 8a for this three time series, where the number of temporal windows is 20 for the first two time series (see Figs. 1, 2) and 30 for the last time series (see Fig. 3). Besides, for the first two time series the number of clusters is set to 3 both for the amplitude and the change of amplitude while the last time series the number of clusters is set to 5 both for the amplitude and the change of amplitude. For each time series, linguistic terms associated with granular prototypes (refer to Figs. 6, 7 and 8) are respectively reported in Tables 1, 3 and 5.

Along with the visualization of the granular prototypes, we provide with their characterization, viz. a linguistic description of each of them and experimental legitimacy level \(\mu \) is reported in Tables 24 and 6. Observe that there is a varying level of legitimacy of the prototypes and some of them could be completely eliminated as not being reflective of any data i.e. the spurious granular prototypes are removed when its value of legitimacy level is zero or close to zero, results of which are represented in Figs. 6b, 7b and 8b.

7 Matching information granules and a realization of linguistic description of time series

For the linguistic description of granular time series, it becomes essential to match the fuzzy relations \(X_i\times \Delta X_i\) (their \(\alpha \)-cuts) with the corresponding prototypes obtained during fuzzy clustering. The essence is to complete matching of \(X_i\times \Delta X_i\) and the prototype \(V_I\times \Delta V_I\) by computing the possibility measure \(Poss(X_i\times \Delta X_i, V_I\times \Delta V_I )\). These calculations are straightforward as we find a maximal value of \(\alpha \) for which the corresponding \(\alpha \)-cuts of \(X_i\times \Delta X_i\) and \(V_I\times \Delta V_I\) overlap. We repeat the calculations for all \(I=1, 2 \dots c\) and choose the index of the prototype \(\lambda _{I0}\) with the highest possibility value. Denote this value by \(\lambda _{I0}\). The process is carried out for successive information granules \(X_i\times \Delta X_i\) thus giving rise to a series of linguistic descriptors along with their levels of matching,

$$\begin{aligned} (V_{I1}\times \Delta V_{I1}, \lambda _1) (V_{I2}\times \Delta V_{I2}, \lambda _2)\dots (V_{Ip}\times \Delta V_{Ip}, \lambda _p)\nonumber \\ \end{aligned}$$
(6)

The details about the realization of the matching process are covered in Appendix C.

Proceeding with the time series discussed so far, in Tabels 7, 8 and 9, we report linguistic descriptions of the time series as a sequence of the linguistic terms, which match the segments of the data to the highest extent. Furthermore we report those linguistic terms whose matching is the second the highest to gain a better sense as to the discriminatory capabilities of the linguistic terms used in the description of the time series.

8 Conclusions

In this study, we have proposed the concept of granular models of time series (granular time series, for short) and elaborated on the ensuing descriptions of the temporal data carried out at the level of information granules. Two conceptual levels of interpretation were clearly identified. It is shown how they endow the model with interpretation capabilities and facilitate interaction with the user. The first one, based on justifiable information granules is oriented towards supporting a graphic vehicle visualizing the nature of time series through a sequence of information granules reported in successive time windows.

In all endeavors, it has to be stressed that information granularity and information granules play a pivotal role by casting all temporal data in a setting, which is easily comprehended by users. The level of detail can be easily adjusted by specifying the number of information granules. The study is focused on the description of time series and the granular models aimed at prediction tasks are only highlighted as they occur at the highest level of the hierarchical structure in Fig. 4. The use of the temporal information granules of variable width (specificity) can bring a unique feature of predictive models when not only the results of prediction are generated but those are supplied with the time horizon over which they are valid.

While the study has offered some conceptual and methodological framework, it is apparent that in the overall discussion information granules are formalized as intervals and fuzzy sets. The principle of justifiable granularity comes with sufficient level of generality where one is able to engage other formalisms of information granularity thus making the overall framework sufficiently general. Furthermore the representation of time series, viz. the feature space could be very diversified. The study used on the simple temporal representation of the time series (information granules built for information granules in the space of amplitude and change of amplitude) however information granules can be equally well formed in some other feature spaces (such as e.g., those resulting from a spectral representation of time series).