1 Introduction

Fuzzy sets were first proposed by Zadeh (1965). Fuzzy sets have been used and are useful in many important studies in the literature. Chen and Chen (2002), and Lin et al. (2006) used fuzzy sets in the construction of fuzzy inference systems as decision support systems. Chen and Lee (2010) utilised fuzzy sets for solving decision-making problems. Chen (2002), Chen and Wang (2010), Chen and Chen (2011, 2014), Chen and Jian (2017), Chen et al. (2019), Zeng et al. (2019), Samal and Dash (2023) and Goyal and Bisht (2023) proposed methods based on fuzzy sets for solving the forecasting problem. Intuitionistic fuzzy sets have an extended definition of fuzzy sets and have an additional dimension for uncertainty. Intuitionistic fuzzy sets are proposed by Atanassov (1986). Chen and Randyanto (2013), Zou et al. (2020) and Meng et al. (2020) contributed and used intuitionistic fuzzy sets literature. In the literature, it has been observed that forecasting methods based on intuitionistic fuzzy sets have been developed and successful forecasting results can be obtained. Although intuitionistic fuzzy time series have different uses in the literature, Egrioglu et al. (2019) defined intuitionistic fuzzy time series in multivariate time series structure. In recent years, many forecasting methods for intuitionistic fuzzy time series have been developed. Abhishekh and Singh (2020) proposed an enhanced and versatile method of forecasting using the concept of intuitionistic fuzzy time series based on their score function. Abhishekh and Singh (2018) presented a refined method of forecasting based on high-order intuitionistic fuzzy time series by transforming historical fuzzy time series data into intuitionistic fuzzy time series data. Xu and Zheng (2019) proposed a long-term forecasting method for intuitionistic fuzzy time series. Kumar et al. (2019) proposed an intuitionistic fuzzy time series forecasting based on a dual hesitant fuzzy set for stock market forecasting. Hassan et al. (2020) designed an intuitionistic fuzzy forecasting model combined with information granules and weighted association reasoning. Wang et al. (2020) proposed a traffic anomaly detection algorithm based on intuitionistic fuzzy time series graph mining. Abhishekh and Singh (2020) proposed a new method of time series forecasting using an intuitionistic fuzzy set based on average length. Fan et al. (2020) proposed a network traffic forecasting model based on long-term intuitionistic fuzzy time series. Pattanayak et al. (2021) proposed a novel probabilistic intuitionistic fuzzy set-based model using a support vector machine for high-order fuzzy time series forecasting.

Chen et al. (2021) proposed an intuitionistic fuzzy time series model based on based on the quantile discretization approach. Kocak et al. (2021) used long short-term memory artificial neural networks to determine the fuzzy relations in their deep intuitionistic fuzzy time series forecasting method. Nik Badrul Alam et al. (2022a) integrated the 4253HT smoother with the intuitionistic fuzzy time series forecasting model. Nik Badrul Alam et al. (2022b) proposed an intuitionistic fuzzy time series forecasting model to forecast Malaysian crude palm oil prices. Arslan and Cagcag Yolcu (2022) proposed an intuitionistic fuzzy time series forecasting model based on a new hybrid sigma-pi neural network. Bas et al. (2022) proposed an intuitionistic fuzzy time series method based on a bootstrapped combined Pi-Sigma artificial neural network and intuitionistic fuzzy c-means. Pant and Kumar (2022a) proposed a novel hybrid forecasting method using particle swarm optimisation and intuitionistic fuzzy sets. Pant and Kumar (2022b) proposed a computational method based on intuitionistic fuzzy sets for forecasting. Pant et al. (2022) proposed a novel intuitionistic fuzzy time series forecasting method to forecast death due to COVID-19 in India. Vamitha and Vanitha (2022) proposed a new model for temperature forecasting using intuitionistic fuzzy time series forecasting. Yolcu and Yolcu (2023) used a cascade forward neural network in the determination of intuitionistic fuzzy relations in their intuitionistic fuzzy time series forecasting model. Dixit and Jain (2023) proposed a new intuitionistic fuzzy time series method for non-stationary time series data with a suitable number of clusters and different window sizes for fuzzy rule generation. Yücesoy et al. (2023) proposed a new intuitionistic fuzzy time series method that uses intuitionistic fuzzy clustering, bagging of decision trees and principal component analysis for forecasting problems. Çakır (2023) proposed a novel Markov-weighted intuitionistic fuzzy time series model for forecasting problems. Kocak et al. (2023) proposed a novel explainable robust high-order intuitionistic fuzzy time series forecasting method based on intuitionistic fuzzy c-means algorithm and robust regression method. Pant et al. (2023) proposed a computational-based partitioning intuitionistic fuzzy time series forecasting method.

When the literature is examined, the need to use both membership and non-membership values leads to the need for more complicated modelling methods in intuitionistic fuzzy time series methods. Another problem is whether the intuitionistic fuzzy time series method should be studied with an approach that uses membership and non-membership values at the same time or with a dual approach with two different models using membership and non-membership values separately and finally a combination of these models. In the literature, the dual approach is generally preferred. The dual approach is based on optimising two different models with two separate and independent optimisation processes. In addition, a third optimisation process is required to combine the outputs of the two models. This leads to the emergence of three different modelling errors. The motivation of this paper is to present an approach that can solve three different modelling error problems and simplification of the model problem. Although intuitionistic fuzzy time series forecasting methods require complicated modelling tools, it is known in the literature that complicated methods may have worse forecasting performance than simple forecasting methods and complicated methods bring memorisation problems. Another motivation for this study is to transform the intuitionistic fuzzy time series method from complicated to simple.

In this study, all weights of two separate dendrite neuron models, one of which works with membership values and the other with non-membership values, are estimated in a single grey wolf optimisation process. In addition, the parameters used in combining the outputs of two dendritic neuron models and combining the final model with exponential smoothing are estimated in the same grey wolf optimisation process. The method proposed in this study combines the output of the intuitionistic fuzzy time series method with the exponential smoothing mechanism, allowing the forecasting method to transform into a simple forecasting method or the method to produce combined forecasts with the forecasts of the simple forecasting method and complicated forecasting method. In this respect, the method has the advantage of having an adaptive and modular structure. The main contribution of this paper is to propose a new forecasting approach that combines the traditional time series forecasting method, the simple exponential smoothing method, with the contemporary forecasting method, the intuitionistic fuzzy time series method.

The other sections of the paper are as follows. In the second part of the study, general information about grey wolf optimisation, intuitionistic fuzzy c-means and the dendrite neuron model is presented. In the third section, the proposed method is introduced. In the fourth section, the application of the proposed method and performance comparison results are explained with the help of tables and graphs.

2 The general information

In this section, summarising information about grey wolf optimisation, intuitionistic fuzzy c-means and dendrite neuron model are presented under subheadings, respectively.

2.1 Grey wolf optimisation algorithm

The grey wolf optimisation algorithm is a meta-heuristic AI optimisation algorithm inspired by the intelligent behaviour of grey wolf packs, proposed by Mirjalili et al. (2014). The method is applied in main steps such as hunting, searching for prey, encircling prey, and attacking prey. The main working logic of grey wolf optimisation is that alpha, beta and delta wolves direct omega wolves according to the order of hierarchy in the hierarchy within the pack and imitate this hierarchical structure. The working logic of the grey wolf optimisation algorithm is given in the following algorithm.

Algorithm 1 Grey wolf optimiser

Step 1. The initial population is created. Uniform distribution is used to generate positions.

Step 2. Alpha, beta, and delta solutions are determined. These are the three best solutions in the population.

Step 3. For each agent in the population, a new agent is created by utilising alpha beta and delta solutions.

Step 4. The population is updated according to whether improvement is achieved with the new solutions obtained. Step 5. Alpha, beta and delta solutions are updated. Here, the previous three best solutions are compared with the new three best solutions.

Step 6. Stop conditions are checked. Steps 3–5 are continued until the stopping conditions are met. If the stop conditions are met, proceed to Step 7. The stop condition is activated when the fitness value does not improve for ten consecutive iterations.

Step 7. The alpha solution is obtained as the solution of the optimisation problem. At this stage, the algorithm reaches the best solution result.

2.2 Intuitionistic fuzzy c-means

Although many different methods are used in clustering data, fuzzy clustering methods offer realistic solutions, especially for the grading of universal cluster elements close to cluster boundaries. In intuitionistic fuzzy clustering, both membership and non-membership values carry two separate important information due to the presence of degrees of hesitation. In the heuristic fuzzy c-means method, iterative equations similar to fuzzy c-means are used to calculate membership values and centres. Let \({u}_{ik}\) denote the membership value of an element of the i-th universal set belonging to the k-th set. Let \({d}_{ik}\) denote the distance value of an element of the i-th universal set belonging to the k-th set. Membership values are calculated with the following formula in intuitionistic fuzzy c-means.

$${u}_{ik}=\frac{1}{{\sum }_{j=1}^{c}{\left(\frac{{d}_{ik}}{{d}_{jk}}\right)}^{2/(m-1)}} \quad i={1,2},\dots ,c;k={1,2},\dots ,n$$
(1)

In Eq. (1), \(c\) is the number of fuzzy sets, \(n\) is the number of observations and m is the fuzziness index. Equation (2) is used in the calculation of hesitation degrees. This equation allows obtaining small hesitation values for membership values close to zero and one, and high hesitation values for membership values close to 0.5.

$${\pi }_{ik} =1-{u}_{ik}-{\left(1-{{u}_{ik}}^{\alpha }\right)}^{1/\alpha }, \alpha >0$$
(2)

The values of \({u}_{ik}^{*}\) given in Eq. (3) indicate the fuzzy membership value consisting of the sum of the heuristic membership and the degree of hesitation.

$${u}_{ik}^{*}={u}_{ik}+{\pi }_{ik}$$
(3)

Fuzzy membership values are used in the calculation of cluster centres with Eq. (4).

$${v}_{i}^{*}=\frac{\sum_{k=1}^{n}{\left({u}_{ik}^{*}\right)}^{m}{x}_{k}}{\sum_{k=1}^{n}{\left({u}_{ik}^{*}\right)}^{m}} ; \quad i={1,2},\dots ,c$$
(4)

In this formula, \({x}_{k}\) denotes the kth data point.

By controlling the magnitude of the change in the membership values, the algorithm of the intuitionistic fuzzy c-means method is obtained with the successive use of the given formulas.

With the application of intuitionistic fuzzy clustering, membership, non-membership and hesitation values of each observation for each cluster can be obtained.

2.3 Dendrite neuron model

In the literature, different models of the biological neuron can be written. A good mathematical model of the biological neuron is the dendrite neuron model. The dendrite neuron model (DNM) was first proposed by Todo et al. (2014). The dendrite neuron model artificial neural network consists of synaptic, dendrite, membrane and soma layers. In the synaptic layer, different non-linear transformations of all inputs are performed. A separate transformation of each input is performed in each dendrite branch.

$${Y}_{ij}=\frac{1}{1+{\text{exp}}(-k\left({w}_{ij}{{\text{Input}}}_{i}+{\theta }_{ij}\right))}$$
(5)

In Eq. (5), index \(i\) indicates the input and \(j\) indicates the dendrtie branch. In the dendrite layer, the inputs for each dendrite are combined with a multiplicative merging function using Eq. (6). Thus, each dendrite branch has an output.

$${Z}_{j}=\left(\prod_{i=1}^{p}{Y}_{ij}\right)$$
(6)

In the membrane layer, the information from the branches is combined with the aggregation function as given in Eq. (7).

$$V=\sum_{j=1}^{M}{Z}_{j}$$
(7)

In the Soma layer, the nonlinear transformation given in Eq. (8) is applied to obtain the output.

$${\text{Output}}=\frac{1}{1+{\text{exp}}\left(-{k}_{{\text{soma}}}(V- {\theta }_{{\text{soma}}})\right)}$$
(8)

Although various training algorithms have been proposed in the literature for training the dendrite neuron model, it is known that artificial intelligence optimisation algorithms can produce successful results for training this network. In the dendrite neuron model, the number of dendrite branches is considered as a hyperparameter and the value of this parameter can be determined by various methods. In the application of artificial intelligence optimisation algorithms, the mean value of the error squares calculated over the training set is preferred as the fitness function.

3 Proposed method

In this study, a new hybrid of intuitionistic fuzzy time series forecasting method and simple exponential smoothing method is proposed. The proposed method can transform into a simple exponential smoothing method by adjusting itself according to the data set. In the proposed method, the time series of interest can be predicted with a complex prediction model or a simple prediction model, or it is possible to realise the weighted prediction of these two approaches. In this respect, the proposed method has a modular and adaptive structure. An important feature of the proposed method is that all parameters in the model can be estimated with a single grey wolf optimiser.

The operation processes of the proposed method are given in Fig. 1. In the proposed method, the real observations of the time series are clustered and membership and non-membership values are obtained for each observation according to the number of predetermined heuristic fuzzy clusters. The outputs of the dendrite neuron models created separately for membership and non-membership values are combined with simple exponential smoothing to obtain the final forecasts. The step-by-step algorithm of the proposed method is given below.

Fig. 1
figure 1

The graphical abstract for the proposed method

Algorithm 2 The proposed method

Step 1. Possible parameter values of the method are determined. These parameters are listed below.

\(c\): The number of intuitionistic fuzzy clusters

\(p\): The number of lagged variables

\(nd\): The number of dendrite branches

\(ntrain\): The length of the training set

\(nval\): The length of the validation set

\(ntest\): The length of the test set

The possible values of these parameters can be determined according to the components of the analysed time series.

Step 2. Time series data are divided into three as training, validation and test data. The last \(ntest\) of the time series is separated as test data and the remaining initial time series is separated as training data as given in Eqs. (9)–(12).

$${x}_{t}=\left[{x}_{1},{x}_{2},\dots .,{x}_{n}\right]$$
(9)
$${x}_{train}=\left[{x}_{1},{x}_{2},\dots ,{x}_{n-ntest-nval}\right]$$
(10)
$${x}_{val}=[{x}_{n-nval-ntest+1}, {x}_{n-nval-ntest+2},\dots ,{x}_{n-ntest}]$$
(11)
$${x}_{test}=[{x}_{n-ntest+1}, {x}_{n-ntest+2},\dots ,{x}_{n}]$$
(12)

The training set is used to determine the optimal weight and side values, the validation set is used to select the hyperparameters, and the test data are used for comparisons with alternative methods.

Step 3. The intuitionistic fuzzy c means method is applied to the training data. The matrices of membership (\(\mu )\) and non-membership \((\gamma\)) values are calculated as given in Eqs. (13)–(14).

$$\mu ={[{\mu }_{ij}]}_{i={1,2},\dots ,n;j={1,2},\dots ,c}$$
(13)
$$\gamma ={[{\gamma }_{ij}]}_{i={1,2},\dots ,n;j={1,2},\dots ,c}$$
(14)

In the application of the heuristic fuzzy c averages method, the fuzziness index (m) is taken as 2 and the alpha as 0.85.

Step 4. Lagged variables are obtained for each column of the matrices of membership and non-membership values. Thus, delayed membership and non-membership matrices that will be input to dendrite neuron models are obtained as given in Eqs. (15)–(16).

$$\mu d={[{\mu d}_{ij}]}_{i={1,2},\dots ,n-p;j={1,2},\dots ,p\times c}$$
(15)
$$\gamma d={[{\gamma d}_{ij}]}_{i={1,2},\dots ,n-p;j={1,2},\dots ,p\times c}$$
(16)

These values are prepared to form inputs to the artificial neural network used.

Step 5. The parameters of the modular neural network architecture given in Fig. 2 are estimated by grey wolf optimisation algorithm. Figure 2, the inputs of the network are the lagged variables of memberships and non-memberships for all clusters.

Fig. 2
figure 2

The architecture of modular and hybrid artificial neural network

The output of the proposed method is given in the following equations. Synaptic layer calculations can be handled in two parts. In Eq. (17), synaptic calculations are performed with membership values, whilst in Eq. (18), synaptic calculations can be performed with non-membership values.

$${Y}_{i,j}^{(d,s,1)}=\frac{1}{1+{\text{exp}}(-k\left({w}_{ij}^{(1)}{\mu }_{i-d,s}+{\theta }_{i,j}^{(1)}\right))}$$
(17)
$${Y}_{i,j}^{(d,s,2)}=\frac{1}{1+{\text{exp}}(-k\left({w}_{ij}^{(2)}{\gamma }_{i-d,s}+{\theta }_{i,j}^{(2)}\right))}$$
(18)

For membership values, the output of the dendrite branch (\({Z}_{j}^{\mu })\) is calculated by Eq. (19) and for non-membership values, the output of the dendrite branch (\({Z}_{j}^{\gamma })\) is calculated by Eq. (20).

$${Z}_{j}^{\mu }=\left(\prod {Y}_{i,j}^{(d,s,1)}\right)$$
(19)
$${Z}_{j}^{\gamma }=\left(\prod {Y}_{i,j}^{(d,s,2)}\right)$$
(20)

The outputs of the membrane layer are similarly calculated by Eqs. (21)–(22) based on membership and non-membership values.

$${V}^{\mu }=\sum_{j=1}^{nd}{Z}_{j}^{\mu }$$
(21)
$${V}^{\gamma }=\sum_{j=1}^{nd}{Z}_{j}^{\gamma }$$
(22)

The outputs of the Soma layer are calculated by Eqs. (23) and (24).

$${{\text{Output}}}^{\mu }=\frac{1}{1+{\text{exp}}\left(-{k}_{{\text{soma}}}^{\mu }({V}^{\mu }- {\theta }_{{\text{soma}}}^{\mu })\right)}$$
(23)
$${{\text{Output}}}^{\gamma }=\frac{1}{1+{\text{exp}}\left(-{k}_{{\text{soma}}}^{\gamma }({V}^{\gamma }- {\theta }_{{\text{soma}}}^{\gamma })\right)}$$
(24)

The final output of the dendrite neural network is A, which is the linear combination of the outputs from the membership and non-membership values. This output is calculated by Eq. (25).

$${{\text{Output}}}^{IFTS}={ w}{{\text{Output}}}^{\mu }+(1-w) {{\text{Output}}}^{\mu }$$
(25)

The output of the proposed method is calculated as the linear combination of the one-step lagged variable and the intuitionistic fuzzy time series part by Eq. (26).

$$\widehat{{x}_{t}}=\left(1-\lambda \right){{\text{Output}}}^{IFTS}+\lambda {x}_{t-1}$$
(26)

The grey wolf optimisation algorithm aims to minimise the mean value of the error squares calculated for the training set. The optimisation problem to be solved with grey wolf optimiser is a constrained multivariate optimisation problem and is shown as given in Eq. (27).

$$\underset{\Xi }{{\text{min}}}\sum \frac{{\left({x}_{t}-\widehat{{x}_{t}}\right)}^{2}}{ntrain}$$
(27)

In Eq. (27), \(\Xi\) denotes all parameters in the proposed method and its elements are as follows.

$$\Xi =[{W}^{\left(1\right)},{W}^{\left(2\right)},{\Theta }^{\left(1\right)},{\Theta }^{\left(2\right)},{k}_{{\text{soma}}}^{\mu },{\theta }_{{\text{soma}}}^{\mu },{k}_{{\text{soma}}}^{\gamma },{\theta }_{{\text{soma}}}^{\gamma },w,\lambda ]$$

\({W}^{\left(1\right)}\),\({W}^{\left(2\right)}\), \({\Theta }^{\left(1\right)}\) and \({\Theta }^{\left(2\right)}\) are \(p\times c\times nd\) dimensional matrices. \({k}_{{\text{soma}}}^{\mu },{\theta }_{{\text{soma}}}^{\mu },{k}_{{\text{soma}}}^{\gamma },{\theta }_{{\text{soma}}}^{\gamma },w\) and \(\lambda\) are scalars. In addition, \(w\) (\(0\le w\le 1\)) and \(\lambda\) \(\left(0\le \lambda \le 1\right)\) may have restricted values in the range between zero and one.

Algorithm 1 is used to solve the optimisation problem (27). In using this algorithm, the values of the solutions corresponding to \(w\) and \(\lambda\) are constrained by Eq. (28).

$$y={\text{min}}\left(1,{\text{max}}\left(0,y\right)\right)$$
(28)

Step 6. In step 4, forecasts for the validation set are obtained using the optimised parameter values. The best hyperparameter values \({c}_{best}\), \({p}_{best}\) and \({nd}_{best}\) are determined based on the validity set performance.

Step 7. The model is retrained using the best hyperparameter values (\({c}_{best}\), \({p}_{best}\) and \({nd}_{best}\)) and the test set forecasts are obtained.

4 Applications

In this study, four time series are used to compare the performance of the proposed method. The time series used in the application are divided into three parts as training-validation and test set. This allows for a more detailed and comprehensive comparison of the methods to be used in the analysis. In order to compare the performance of the proposed method, the Chen (2002) high-order fuzzy time series solution method, which is one of the classical fuzzy time series methods, the fuzzy time series method based on the multiplicative neuron model (FTS-SMN) proposed in Aladağ (2013), the picture fuzzy regression functions method (PFF) proposed in Bas et al. (2020), and the high-order univariate heuristic fuzzy time series prediction model (IFTS) method in Egrioglu et al. (2019) were used.

The first time series used in the application is the time series of the Index of Coin Market Cap opening prices (CMC-OPEN) between 12/07/2022 and 13/02/2023. The time series contains a total of 150 observations. The graph of the time series is given in Fig. 3. The time series is downloaded from the Yahoo Finance website (https://finance.yahoo.com/).

Fig. 3
figure 3

CMC OPEN 150 opening prices between 12/07/2022 and 13/02/2023

The results obtained by analysing the time series are given in Table 1. Table 1 presents the mean and standard deviation of the root of mean square error values for 30 different random initial weights. The results for the Chen (2002) method and the best hyperparameter setting are also given in Table 1.

Table 1 Implementation results for CMC-OPEN between 12/07/2022 and 13/02/2023

In Table 1 and all other Tables, \({c}_{best}\) is the best number of clusters, \({p}_{best}\) is the best number of lagged variables and \({nd}_{best}\) is the best number of dendrite neurons. Also, (–) indicates that the method does not have a corresponding value for a cell of the table.

Table 1 shows that the mean statistics of the proposed method is smaller than all other methods

The second time series used in the application is the time series of the opening prices of CMC-OPEN between 5/04/2022 and 7/11/2022. The time series contains a total of 150 observations. The graph of the time series is given in Fig. 4. The time series is downloaded from the Yahoo Finance Web site (https://finance.yahoo.com/).

Fig. 4
figure 4

CMC-OPEN opening prices between 5/04/2022 and 7/11/2023

The results obtained by analysing the time series are given in Table 2.

Table 2 Implementation results for CMC-OPEN between 5/04/2022 and 7/11/2023

Table 2 shows that the proposed method is the second-best method after Aladağ (2013)

The third time series used in the application is the time series of the opening prices of CMC-OPEN between 11/08/2022 and 16/03/2023. The time series contains a total of 150 observations. The graph of the time series is given in Fig. 5. The time series is downloaded from the Yahoo Finance Web site (https://finance.yahoo.com/).

Fig. 5
figure 5

CMC-OPEN opening prices between 11/08/2022 and 16/03/2023

The results obtained by analysing the time series are given in Table 3.

Table 3 Implementation results for CMC-OPEN between 11/08/2022 and 16/03/2023

Table 3 shows that the mean statistics of the proposed method are smaller than all other methods

Finally, the time series of CMC-OPEN opening prices between 16/11/2022 and 23/06/2023. The time series contains a total of 150 observations. The graph of the time series is given in Fig. 6. The time series is downloaded from the Yahoo Finance website (https://finance.yahoo.com/).

Fig. 6
figure 6

CMC-OPEN between 16/11/2022 and 23/06/2023

The results obtained by analysing the time series are given in Table 4.

Table 4 Implementation results for CMC-OPEN between 16/11/2022 and 23/06/2023

Table 4 shows that the mean statistics of the proposed method are smaller than all other methods

5 Conclusions

In this study, for the first time in the literature, a new forecasting method is proposed by integrating the intuitionistic fuzzy time series method with simple exponential smoothing. The dendrite neuron model artificial neural network, which works with delayed values of membership and non-membership values, is hybridised with exponential smoothing and a new architectural structure is presented. The main contribution of this paper is to introduce a new combined forecasting method to the literature.

The proposed method can balance between two simple and two complex forecasting models and can automatically increase the bias towards either one. It has been observed that the proposed method can produce more successful prediction results than the current prediction methods in the literature. Future studies may focus on the development of integrated prediction systems with different exponential smoothing methods. In addition, it is an important research problem to address the method in the concept of explainable artificial intelligence.