Journal of Big Data

, 5:49 | Cite as

Building efficient fuzzy regression trees for large scale and high dimensional problems

  • Javier CózarEmail author
  • Francesco Marcelloni
  • José A. Gámez
  • Luis de la Ossa
Open Access


Regression trees (RTs) are simple, but powerful models, which have been widely used in the last decades in different scopes. Fuzzy RTs (FRTs) add fuzziness to RTs with the aim of dealing with uncertain environments. Most of the FRT learning approaches proposed in the literature aim to improve the accuracy, measured in terms of mean squared error, and often neglect to consider the computation time and/or the memory requirements. In today’s application domains, which require the management of huge amounts of data, this carelessness can strongly limit their use. In this paper, we propose a distributed FRT (DFRT) learning scheme for generating binary RTs from big datasets, that is based on the MapReduce paradigm. We have designed and implemented the scheme on the Apache Spark framework. We have used eight real-world and four synthetic datasets for evaluating its performance, in terms of mean squared error, computation time and scalability. As a baseline, we have compared the results with the distributed RT (DRT) and the Distributed Random Forest (DRF) available in the Spark MLlib library. Results show that our DFRT scales similarly to DRT and better than DRF. Regarding the performance, DFRT generalizes much better than DRT and similarly to DRF.


Fuzzy regression trees Big Data Fuzzy discretizer Apache Spark 





decision tree


distributed fuzzy regression tree


Distributed Random Forest


distributed regression tree


fuzzy regression tree


Hadoop Distributed FileSystem


Map Reduce


mean squared error


partitioning based on fuzzy variance


Random Forest


regression tree


Regression trees (RTs) are light but powerful models that have been extensively used in different application domains [13, 24, 35, 37]. Their main advantage relies on their simplicity both in the learning and execution phases. Furthermore, RTs are highly interpretable models, that is, they can be used to describe the relation between the inputs and the output. An RT is a tree-shaped directed graph in which each internal (non-leaf) node represents a test on an attribute and each leaf node holds an output value. Each branch, which consists of a sequence of non-leaf nodes (tests) and a leaf node (output value), represents a possible outcome for such group of tests. By following the corresponding branches from the root node to a leaf, one can find an explanation of how the output has been determined. In the literature, several different algorithms for learning RTs have been proposed [5, 20, 36].

The type of tests used in the internal nodes determines the partitioning of the input space. Given a parent node, the test acts as a splitting method for generating its child nodes. There are two main categories of RTs: binary and multi-way RTs. Binary RTs recursively split the input space into two subspaces for each test node, so that each parent node is connected exactly to two child nodes. On the other hand, multi-way RTs split the input space into k partitions, where \(k>1\). Therefore, each parent node is connected to a not fixed number of child nodes. The structure of a multi-way RT is usually more compact and interpretable than the structure of a binary RT [3, 22]. On the other hand, multi-way splits lead to more accurate RTs but, since they tend to fragment the training data very quickly [19], they generally need larger training datasets to work effectively [22]. Furthermore, a multi-way RT can be also represented as a binary RT [14].

One of the main drawbacks of RTs is related to the crisp bounds of the branch conditions: a small change in the values of the input variables may produce an important difference in the prediction. To overcome this problem, RTs have been extended with the use of fuzzy set theory, originating fuzzy regression trees (FRTs) [31, 39]. FRTs employ fuzzy conditions in their internal nodes. Therefore, an input can enable several branches in a test node, and reach different leaves, with different confidence levels. The output value of the FRT is obtained by appropriately combining the values associated with these multiple leaf nodes. FRTs are, in general, more complex both in learning and execution phases than RTs.

Fuzzy regression tree learning algorithms usually apply an initial discretization step aimed at generating fuzzy partitions on the continuous input variables, typically guided by some purposely-defined index [43, 47]. Since this step has a direct impact on the performance of the learning algorithm, several studies evaluate how the accuracy and complexity of the generated models depend on the discretization [15, 23, 50].

Most of FRT learning algorithms proposed in the literature focus on the accuracy but do not consider the computation complexity and space requirements [33, 39]. Typically, these algorithms have been tested and assessed on small datasets, and are not generally suitable for managing huge amounts of data. In fact, one solution often adopted to work in real contexts with these algorithms, is to reduce the dataset by applying downsampling techniques. However, these techniques may cut off some useful knowledge, making FRT learning approaches purposely designed for managing the overall dataset more desirable and effective. Thus, we aim to learn FRTs from the overall dataset, independently of its size. In other words, we desire to cope explicitly with Big Data.

As described in [42], Big data refers to the storage and analysis of large and/or complex data sets using a series of specific techniques. These techniques combine new paradigms (software) together with specific hardware architectures. A recent work [38] of one of the co-authors has proposed a novel fuzzy decision tree algorithm to address Big Data classification problems. First, the domain of the continuous input features is discretized by adopting a distributed approach based on the fuzzy entropy concept [50]. Then, the fuzzy decision tree is learned by applying a distributed learning algorithm based on the fuzzy information gain concept. Both phases have been implemented by using the MapReduce paradigm in the Apache Spark framework.

In this paper, we adapt the algorithm proposed in [38] to deal with Big Data regression problems. In particular, we adopt more suitable metrics than fuzzy entropy and information gain for the discretization and FRT learning phases. Furthermore, we employ some specific arrangements for speeding-up the distributed implementation of the FRT. Finally, we perform an extensive experimentation of our FRT learning algorithm by using eight real-world datasets [2] and four synthetic datasets [40]. The results are discussed in terms of mean squared error, computation time and scalability, and compared with the ones achieved by the distributed regression tree and the Distributed Random Forest [4], both included in the MLlib Apache Spark library [30]. To the best of our knowledge, the proposed algorithm is the first distributed FRT learning scheme proposed in the literature.

The paper is organized as follows. “Background” section introduces some preliminary concepts about RTs and FRTs. Afterwards, in “MapReduce and spark” section, we briefly discuss MapReduce and Apache Spark. “Fuzzy regression tree for Big Data” section describes our proposed FRT learning algorithm and presents its distributed implementation. Then, in “Results and discussion” section, we present the experimental results. Finally, in “Conclusions” section, we draw our conclusions.


In classification problems, a class \(c_m\) from a predefined set \({\mathcal {C}}= \{c_1, \ldots , c_M\}\) of M classes is assigned to an unlabeled instance. A classification problem is defined by a set of input variables \(\mathbf X = \{ X_1, \dots , X_F \}\) and the output Y. In case of numerical variables, \(X_f\) is defined on a universe \(U_f\subset \mathfrak {R}\). In case of categorical variables, \(X_f\) is defined on a set \(L_{f} = \{L_{f,1}, \dots , L_{f,T_{f}} \}\) of categorical values.

Decision trees (DTs) have been widely used in instance classification problems [11, 16, 52]. A DT is a tree-shaped directed graph in which each internal (non-leaf) node represents a test on an attribute. Each path from the root to a leaf corresponds to a sequence of tests, which aims to isolate a subspace of the input space. The objective of the test is to partition the training set into subsets as pure as possible, that is, subsets composed by instances belonging to the same class. In the case of numerical variables, tests are defined by intervals in the definition domain; in the case of categorical variables, they correspond to subsets of the possible categorical values. Formally, tests are in the form of \(X_f > x_{f,s}\) and \(X_f \le x_{f,s}\) for numerical input variables, and \(X_f \subseteq {\mathbf {L}}_{f,s}\) for categorical input variables, where \(x_{f,s}\) and \({\mathbf {L}}_{f,s}\) are a numeric threshold and a set of categorical values for the test s, respectively. Each leaf node is characterized by a class \(c_m \in \mathcal {C}:\) the instances, which satisfy the overall sequence of tests from the root to the leaf, are classified as belonging to \(c_m.\)

Decision trees can also be used in regression problems, named regression trees (RTs). The objective is to predict a real value rather than a class label. Thus, leaf nodes are characterized by a regression model defined over the input variables. For instance M5 [36] generates first-order polynomials using the overall set of input variables. CART [5], which is one of the most known algorithms for generating regression trees, uses zero-order polynomials i.e. a numerical constant, leading to simpler and usually more robust models. Indeed, the use of high order polynomials tends to overfit.

In this work, we focus on zero-order polynomial fuzzy regression trees (FRTs) [18]. Each real input variable is partitioned by using fuzzy sets. The tests in the internal nodes use these fuzzy sets in the form of \(X_f\) is \(B_{f,t},\) where \(B_{f,t}\) is a fuzzy set defined over the variable \(X_f.\) The membership degree of a value \(x_{i,f}\) with respect to the fuzzy set \(B_{f,t}\) is represented as \(\mu _{B_{f,t}}(x_{i,f}).\) Similarly, the membership degree of an instance \(\mathbf x _{i}\) with respect to the fuzzy sets \(B_{1,t_1}, \ldots , B_{F,t_F}\) is represented as T-Norm\(\left(B_{1,t_1}(\mathbf x _{i}), \ldots , B_{F,t_F}(\mathbf x _{i})\right),\) which is a conjunction of the individual membership degrees.1 Since fuzzy sets generally overlap, an input instance may activate more than one leaf node. The value assigned to each leaf node is computed as a weighted average of the output values for all the training set instances that activate such leaf node, using the activation degrees as instance weights.

There are different approaches for generating the fuzzy partitions of the input variables automatically from data [23, 27, 50]. Most of them use two steps: discretization and fuzzy partition generation. The first step consists of discretizing the domain of each input variable into a finite set of disjoint bins. To this aim, different strategies have been proposed in the literature [17]: splitting vs. merging, supervised vs. unsupervised, dynamic vs. static, and global vs. local. Once the bins have been generated, in the second step the membership functions of the fuzzy partition are derived from these bins by employing different methods: the use of the bin bounds to define the fuzzy sets sequentially ordered [34], based on histograms [29], probabilities [12], fuzzy nearest neighbors [21], neural networks [29], entropy [6], particle swarm optimization [32] or clustering based methods [29]. In this paper, we adopt an approach that generates triangular fuzzy partitions directly by using a single step.

Algorithm 1 shows the scheme of a generic FRT learning process, where \(\mathbf {TR} = \{(\mathbf {x}_1, y_1), \ldots , (\mathbf {x}_N, y_N)\}\) is the training set, \(\mathbf {X}\) is the set of input variables, SplitMet is the splitting method, StopMet is the stopping method, and FRT is the output (learnt model).

The algorithm first creates the root node. Then, it builds the FRT model with the TreeGrowing recursive function. This function first checks if the node is a leaf by means of the StopMet function. If not, it calls the function SelectBestFeatureSplits, which selects the best input variable to partition the input space. There are several metrics that can be used to perform the variable selection. However, the most popular one in zero-order FRTs is the variance [7, 10]. Finally, for each \(split_z\) in splits the TreeGrowing function creates a child node and recursively calls itself using the subset of instances, which satisfy the test for such child node.

After learning the tree structure, each leaf node \(n_i\) in the set of leaf nodes, LN, estimates a regression model from the instances in the training set, which activate such leaf node, i.e., which satisfy the sequence of tests from the root to such leaf node. Let \(S_{n_i}\) be the set of instances that activate the leaf node \(n_i\). \(RM_{n_i}\) is the weighted average of the instances \((x_l, y_l) \in S_{n_i}\) by using their activation degree \(\mu _{n_i^l}({\mathbf {x}}_l)\) as weight:
$$\begin{aligned} RM_{n_i} = \frac{\sum _{(x_l, y_l) \in S_{n_i}} y_l \cdot \mu _{n_i^l}({\mathbf {x}}_l) }{ \sum _{n_i \in LN} \sum _{(x_l, y_l) \in S_{n_i}} \mu _{n_i^l}({\mathbf {x}}_l) } \end{aligned}$$
During the inference, the prediction is computed as follows. Let \(N^l=\left\{ n_1^l,\ldots ,n_{M^l} \right\}\) be the set of leaf nodes activated by the input instance \({\mathbf {x}}_l\) with an activation degree \(\mu _{n_i^l}({\mathbf {x}}_l)\). The output of the FRT is obtained as:
$$\begin{aligned} \text{ FRT }(\mathbf {x}_l) = \frac{\sum _{n_i \in \mathbf {N^l}} RM_{n_i} \cdot \mu _{n_i}({\mathbf {x}}_l) }{\sum _{n_i \in \mathbf {N^l}} \cdot \mu _{n_i}({\mathbf {x}}_l)} \end{aligned}$$

MapReduce and Spark

MapReduce (MR) is a programming paradigm proposed in 2004 [9] for distributing the computation of parallel data-oriented tasks across a cluster of machines. At the beginning, it was developed as a solution to improve the performance of the Google Search Engine. Currently, its usage has been extended to process large amounts of data in multiple different areas [25].

MapReduce presents an abstraction layer to the developers: It divides the problem into smaller tasks, called Map and Reduce, and executes those tasks in parallel taking care of communication, network bandwidth, disk usage and possible failures. In this way, the developers are able to implement parallel algorithms by simply defining the Map and Reduce functions, avoiding problems related to the underlying architecture and hardware configuration. In order to execute these tasks, MR follows a master-slave scheme. First, the data is automatically divided into a set of independent blocks, called chunks, which can be processed in parallel by different machines [26]. Then, the master computing node configures multiple tasks, each one fed with a chunk of data, and executes them in parallel in slave computing nodes.

The data managed by the MR workflow are represented by \(\langle key, value\rangle\) pairs. Initially, an arbitrary key is assigned to each input. Then, in the Map phase, each task processes the data by applying the defined map function, which generates a list of intermediate \(\langle key, value\rangle\) pairs (one per map task). Then, the system collects these pairs, sort them according to their keys, and feed reduce tasks with the pairs sharing the same key. Finally, in the Reduce phase, the defined reduce function is used to combine the list of values (which share the same key) to produce a new \(\langle key, value\rangle\) pair. The result of the MapReduce process is a list of \(\langle\)key, value\(\rangle\) pairs generated by all the executed Reduce tasks.

The most extended implementation of this paradigm is Apache Hadoop [45], where the open-source community worked together with many private companies to provide a general framework for Big Data processing. The key parts of Apache Hadoop are the MapReduce implementation and the Hadoop Distributed FileSystem (HDFS), which allows working efficiently with large amount of data following the previously described processing.

The most important drawback of Apache Hadoop is that the output is written in the hard driven storage. This decreases the performance for iterative (or stream-based) workflows. Apache Spark [49] is a framework, which overcomes this problem by storing the data structures in the main memory (significantly faster) until the whole process finishes (the process can be a sequence of MapReduce tasks). In practice, the main difference with respect to Apache Hadoop is the resilient distributed dataset (RDD), which relax the MapReduce workflow and allows the developer to define and use additional operations over the data.

Spark has been implemented using the Scala programming language, but there are alternative languages, which can be used for developers, like Python and R. The existence of such interfaces eases its usage and has made Apache Spark one of the most popular frameworks for Big Data processing, in particular within the Data Science community. As a consequence, MLlib [48] has been developed as part of the Apache Spark project. MLlib is a library for Apache Spark, which implements the most popular machine learning algorithms following the MapReduce paradigm.

Fuzzy regression tree for Big Data

In this section, we present our FRT learning approach for big data regression problems. We have adapted the distributed fuzzy decision tree learning approach, proposed by one of the authors in [38] for handling Big Data classification problems, to deal with regression problems. The learning approach consists of two steps: the first one creates fuzzy partitions on the domains of the input variables, and the second one induces a binary FRT from data using the fuzzy partitions generated in the first step. With respect to the original approach, we have employed different measures for both evaluating the partitions in fuzzy partitioning and for determining the optimal variable to be used in a decision node in the FRT learning phases. Consequently, we have had to introduce different stopping conditions. Further, we have employed some specific adjustments for speeding-up the FRT generation.

In the following, we first describe both steps and then we explain how these steps have been designed to be implemented under the MapReduce paradigm.

Fuzzy partitioning

Partitioning of continuous variables is a crucial aspect in the generation of both FDTs and FRTs, and therefore should be performed carefully. The work in [50] evaluates several approaches for FDTs. Among them, fuzzy partitioning based on fuzzy entropy proved to be very effective [38]. Most of these metrics are not suitable for fuzzy partitioning in regression problems, which requires specific indexes for measuring the quality of the partition. In regression, one of the most used indexes is fuzzy variance [8]. Thus, we have adapted the recursive supervised fuzzy partitioning method proposed in [38] for classification problems, by using fuzzy variance as metric. The adapted version, denoted as partitioning based on fuzzy variance (PFV), is a recursive supervised method, which generates candidate fuzzy partitions, and evaluates these partitions employing fuzzy variance rather than the fuzzy entropy as used in the classification-oriented method.

Let \(TR_f = [(x_{1,f}, y_1), \ldots , (x_{N,f}, y_N)]\) be the projection of the training set \(\mathbf {TR}\) along the input variable \(X_f.\) Let us assume that the values \(x_{i,f}\) are sorted in increasing order. Let \(I_{f}\) be an interval defined on the universe of \(X_f.\) Let \(l_f\) and \(u_f\) be the lower and upper bounds of \(I_f.\) Let \(S_f\) be the set of values \((x_{i,f},y_i) \in TR_f\) contained in \(I_f.\) Let \(P_{I_f}=\{B_{f,1},\ldots ,{B_{f,K_{P_{I_f}}}\}}\) be a fuzzy partition defined on \(I_f,\) where \(K_{P_{I_f}}\) is the number of fuzzy sets in \(P_{I_f}.\) Let \(S_{f,1},\ldots ,S_{f,K_{P_{I_f}}}\) be the subsets of points in \(S_f,\) contained in the support of \(B_{f,1},\ldots ,{B_{f,K_{P_{I_f}}},}\) respectively.

The fuzzy mean \(\left(\text{ FM }({P_{I_f}})\right)\) of a fuzzy partition \(P_{I_f}\) is defined as the mean of the output values \(y_i\), corresponding to the instances whose value \(x_{i,f} \in S_f\), weighted by the membership degrees of \(x_{i,f}\) to any \(B_{f,j} \in P_{I_f}\):
$$\begin{aligned} \text{ FM }({P_{I_f}}) = \frac{\sum ^{K_{P_{I_f}}}_{j=1} \sum _{(x_{i,f},y_i) \in S_{f,j}} y_i \cdot \mu _{B_{f,i}}(x_{j,f})}{\sum ^{K_{P_{I_f}}}_{j=1} \sum _{(x_{j,f},y_i) \in S_{f,j}} \mu _{B_{f,j}}(x_{i,f})} \end{aligned}$$
The fuzzy variance \(\left(\text{ FVar }(P_{I_f})\right)\) of a fuzzy partition \(P_{I_f}\) is defined as:
$$\begin{aligned}\text{ FVar }({P_{I_f}}) = \frac{\sum ^{K_{P_{I_f}}}_{j=1} \sum _{(x_{i,f},y_i) \in S_{f,j}} (y_i - \text{ FM }({P_{I_f}}))^2 \cdot \mu ^2_{B_{f,j}}(x_{i,f}) }{ \sum ^{K_{P_{I_f}}}_{j=1} \sum _{(x_{i,f},y_i) \in S_{f,j}} \mu ^2_{B_{f,j}}(x_{i,f}) } \end{aligned}$$
At the beginning of the execution of PFV, \(I_f\) coincides with the universe of \(X_f\) and \(S_f=TR_{f}.\) First of all, PFV sorts the values of \(TR_{f}\) in ascending order. Afterward, for each value \(x_{i,f}\) between \(l_f\) and \(u_f\) (at the beginning of the partitioning procedure, \(i=1,\ldots ,N\)), PFV defines a strong fuzzy partition \(P_{I_f}(x_{i,f})\) on \(I_f\) by using three triangular fuzzy sets, namely \(B_{f,1},\) \(B_{f,2}\) and \(B_{f,3},\) as shown in Fig. 1. The cores of \(B_{f,1},\) \(B_{f,2}\) and \(B_{f,3}\) coincide with \(l_f,\) \(x_{i,f}\) and \(u_f,\) respectively. PFV computes \(FVar(P_{I_f}(x_{i,f}))\) for each fuzzy partition \(P_{I_f}(x_{i,f})\) induced by each \(x_{i,f}\) and selects the optimal value \(x^{0}_{i,f},\) which minimizes \(FVar(P_{I_f}(x_{i,f})).\)
This value identifies the fuzzy partition \(P^0_{I_f}(x^0_{i,f})=\{B^{0}_{f,1}, B^{0}_{f,2}, B^{0}_{f,3}\}.\) Let \(S^0_{f,1},\) \(S^0_{f,2}\) and \(S^0_{f,3}\) be the subsets of points in \(S_f,\) contained in the support of the three fuzzy sets, respectively. Then, PFV applies recursively the procedure for determining the optimal strong fuzzy partition to the intervals \(I_f=I^0_{f,1} =[l_f,x^{0}_{i,f}]\) and \(I_f=I^0_{f,2}=(x^{0}_{i,f}, u_f]\) identified by \(x^{0}_{i,f},\) by considering \(S_f=S^0_{f,1}\) and \(S_f=S^0_{f,3},\) respectively, until the following stopping conditions are met:
  • the cardinality \(\vert S_f \vert\) of \(S_f\) is lower than a fixed threshold \(\lambda _{\text{disc}}\)

  • the partition fuzzy gain (\(\text{ PFGain }(P_{I_f})\)) is lower than a fixed threshold \(\epsilon _{\text{disc}}\)

\(\text{ PFGain }(P_{I_f})\) measures the improvement in terms of fuzzy variance of the involved intervals. It is defined as follows:
$$\begin{aligned} \text{ PFGain }(P_{I_f}) &= FVar(P_{I_f}) -FVar(P_{I_{f,1}}) \cdot w(P_{I_{f,1}}) \\&\quad - FVar(P_{I_{f,2}}) \cdot w(P_{I_{f,2}}) \end{aligned}$$
where \(w(P_{I_{f,s}}),\) with \(s=1,2,\) is defined as:
$$\begin{aligned} w(P_{I_{f,s}}) = \frac{\sum ^{K_{P_{I_{f,s}}}}_{g=1} \sum _{(\mathbf {x}_{i,f},y_i) \in S_{f,s}} \mu ^2_{B_{f,g}}(\mathbf {x}_{i,f})}{ \sum ^{K_{P_{I_f}}}_{j=1} \sum _{(\mathbf {x}_{i,f},y_i) \in S_f} \mu ^2_{B_{f,j}}(\mathbf {x}_{i,f}) } \end{aligned}$$
Fig. 1

An example of the fuzzy partition \(P_{I_f}(x_{i,f})\) in the interval \(I_{f}\) defined by \(l_f,\) \(x_{i,f}\) and \(u_f\)

If no initial partition exists on \(I_f\) (this occurs when \(I_f\) coincides with the universe of \(X_f\) and \(S_f = TR_f\)), we assume two fuzzy sets \(B_{f,1}\) and \(B_{f,2}\) defined on \(I_f,\) as shown in Fig. 2a. Once the fuzzy partition with three fuzzy sets is accepted, the procedure generates a partition with three fuzzy sets from intervals \(I_{f,1}\) and \(I_{f,2}\) as shown in Fig. 2b. The partitioning of both \(I_{f,1}\) and \(I_{f,2}\) may generate three fuzzy sets in both \([l_f,x_{i,f}]\) and \((x_{i,f},u_f].\) Actually, the two fuzzy sets, which have the core in \(x_{i,f}\) are fused generating a unique fuzzy set. This fusion can be applied at each level of the recursion. The final result is a strong triangular fuzzy partition \(P_f=\{A_{f,1},\ldots ,A_{f,T_f}\}\) on \(U_f,\) where \(A_{f,j}\) is the j-th fuzzy set defined on \(U_f.\)
Fig. 2

An example of the iterative partitioning process

FRT learning

Once fuzzy partitions have been defined, the pseudo-code described in Fig. 1 is employed to build the FRT. In the SelectBestFeatureSplits function, we use the expression \(\text{ PFGain }(P_{I_f}),\) defined in Eq. 4, to determine the best binary split. To calculate the split with the maximum \(\text{ PFGain }(P_{I_f}),\) we evaluate all possible candidates, by grouping together adjacent fuzzy sets into two disjoint groups \(Z_{1}\) and \(Z_{2}\). The two subsets \(G_{1}\) and \(G_{2}\) of instances contain the points that belong to the support of the fuzzy sets contained in \(Z_1\) and \(Z_2\), respectively. A fuzzy partition with \(T_{f}\) fuzzy sets generates \(T_{f}-1\) candidates. Starting with \(Z_1=\{A_{f,1}\}\) and \(Z_2=\{A_{f,2},...,A_{f,T_{f}}\}\), we compute the fuzzy gain by applying Eq. 4, with \(P_f=\{Z_1,Z_2\}\) and cardinality \(\left| G_1\right| = \sum _{i=1}^{N_1}TN(\mu _{A_{f,1}}(x_{f,i}), \mu _{G}(\mathbf x _{i}))\) and \(\left| G_2\right| = \sum _{i=1}^{N_2}TN(\mu _{A_{f,2}}(x_{f,i})+ \cdots + \mu _{A_{f,T_f}}(x_{f,i}), \mu _{G}(\mathbf x _{i}))\), where \(N_1\) and \(N_2\) are the numbers of instances in the support of the fuzzy sets in \(Z_{1}\) and \(Z_{2}\), respectively, and \(\mu _{G}(\mathbf x _{i})\) is the membership degree of instance \(\mathbf x _{i}\) to the parent node. Iteratively, the algorithm investigates all candidates by moving the first fuzzy set in \(Z_2\) to \(Z_{1}\) and computing the corresponding \(\text{ PFGain }(P_{I_f})\), until \(Z_{2}=\{A_{f,T_{f}}\}\). The pair (\(Z_{1},\,Z_2\)), which obtains the highest \(\text{ PFGain }(P_{I_f})\), is used to create the two child nodes. The two nodes contain, respectively, the examples that belong to the support of the fuzzy sets in \(Z_1\) and \(Z_2\).

In case of categorical variables the computation cost can become very prohibitive: a categorical variable with L values generates \(2^{L-1}-1\) possible combinations. We have adopted the same approach used in [38]: we reduce the possible combinations by sorting the categorical values according to an arbitrary order (by appearance in the dataset). Afterward, we use each categorical value \(x_{j,f}\) as splitting point: the first subset is composed of all the values \(x_{i,f}\) such that \(i \le j,\) leaving the rest for the second subset. Therefore, the number of possible partitions is reduced to \(L-1.\)

The process terminates when one of the following conditions is met (StopMet procedure):
  • The number of instances that reach the node is smaller than a fixed threshold \(\lambda _{\text{learn}}\).

  • The \(\text{ PFGain }(P_{I_f})\) is lower than a threshold \(\epsilon _{\text{learn}}\).

  • The tree has reached a maximum fixed depth \(\beta\).

The output of the model is computed as in Eq. 1. As we mentioned above, in the case of binary FRTs, an input variable can be considered more than once in the same path from the root to a leaf node, as the partitioning is done in a recursive way where an interval \(I_f\) can be split into subintervals at different levels. The use of the T-Norm tends to penalize the matching degree of the instance for those input variables that are repeatedly selected along a same path (from the root to a leaf node). To reduce this effect, we use a strategy that remembers each fuzzy set used in the path, and its membership degree is computed and used only the first time it appears; for the subsequent times the membership degree used is 1.

Distributed approach

We have adapted the distributed implementation proposed in [38] for learning distributed fuzzy decision trees, based on the Map-Reduce paradigm and implemented in Apache Spark. Let V be the number of chunks used for splitting the training set and Q the number of computing units (CUs). Each chunk feds only one Map task, while one CU can process several tasks, both Map and Reduce. Obviously, only Q tasks can be executed in parallel.

The distributed approach follows the same procedure described in the previous sections: first it generates the fuzzy partition of each input variable and then builds the distributed FRT, denoted as DFRT in the following.

With respect to the fuzzy partitioning, we use an approximation of the PFV approach described in “Fuzzy partitioning” section. Indeed, in case of big data, the number of possible candidate splitting points, corresponding to the different values \(x_{i,f}\) in the training set for each input variable \(X_f\), might lead to intractable computation times. To reduce this problem, we limit the number of possible candidate splitting points by adopting an equi-frequency discretization of the domains of each input variable. In particular, as already used in [38], for each input variable \(X_f\), we sort the values \(x_{i,f}\) in the training set and partition the domain of \(X_f\) into a fixed number L of equi-frequency bins. Then, we aggregate the lists of the bin boundaries generated for each chunk and, for each pair of consecutive bin boundaries, we generate a new bin and compute the required statistics for this bin to be used for computing the fuzzy variance at each iteration of PFV. PFV applies the same process described in “Fuzzy partitioning” section, using the central values of such bins as candidate splitting points. Figure 3 shows the distributed PFV.
Fig. 3

Distributed partitioning based on fuzzy variance

In the first phase, distributed bin generation, the Map-Reduce implementation of PFV scans the training set to generate at most \(V \cdot L\) bin boundaries. Each Map task loads a chunk v of data and, for each continuous input variable \(X_f\), sorts its values and computes L equi-frequency bin boundaries, including the left and and right extremes of the \(X_f\) domain in the chunk. Let \(BB_{v,f} = \{b^{(1)}_{v,f}, \ldots , b^{L}_{v,f}\}\) be the sorted list of bin boundaries. The output of the Map task is a key-value pair \(\langle key = f, value = BB_{v,f}\rangle\), where f is the index of the input variable \(X_f\) and \(BB_{v,f}\) is the sorted list of bin boundaries. Afterwards, each Reduce task is fed by V lists of bin boundaries for the same input variable \(X_f\), and outputs a key-value pair \(\langle key = f, value = BB_{f}\rangle\), where \(BB_{f}\) is the sorted list of the bin boundaries for that variable \(X_f\) (obtained by joining all the bin boundaries of each \(BB_{v,f}\)). If the number of bins is equal to the number of values in the interval (one value per bin), it is equivalent to PFV. On the other side, the lower the value of L is, the coarser the approximation in determining the fuzzy partition is. We choose L equal to the percentage \(\gamma _{\text{disc}}\) of the chunk size.

In the second phase, distributed fuzzy sets generation, we generate the fuzzy partitions for each continuous input variable. Each Map task is fed by the bin boundaries list. Let \(x_{i,f} \in S^{(t)}_{f,v}\) be the instances in the chunk v that are contained in the bin \(b^{(t)}_{f}.\) Then, for each input variable \(X_f\) and for each bin \(b^{(t)}_{f}\) in \(BB_{f},\) the Map task computes three statistics \(\mathbf {\varOmega ^{t,v,f}} = \{\varOmega _1^{t,v,f}, \varOmega _2^{t,v,f}, \varOmega _3^{t,v,f}\},\) where
  • \(\varOmega _1^{t,v,f}\) is the sum of the outputs \(y_i\) corresponding to \(x_{i,f} \in S^{(t)}_{f,v}\);

  • \(\varOmega _2^{t,v,f}\) is the sum of the squared outputs \(y_i^2\) corresponding to \(x_{i,f} \in S^{(t)}_{f,v}\);

  • \(\varOmega _3^{t,v,f}\) is the number of instances \(x_{i,f} \in S^{(t)}_{f,v}\).

The output of the Map phase is a list of pairs \(\langle key = f, value = \mathbf {\varOmega ^{t,v,f}} \rangle\). Each Reduce task is fed by a list of V statistics \(\mathbf {\varOmega ^{t,v,f}}\) for the input variable \(X_f\). In a first phase, the Reduce task performs an element-wise addition for all the three statistics, obtaining \(\mathbf {\varOmega ^{v,f}}\). Finally, it applies the PFV described in “Fuzzy partitioning” section. Let \(\overline{b}^{(t)}_{f}\) be the central point of each bin \(b^{(t)}_{f}\). In order to compute efficiently the PFGain(\(P_{I_f}\)) given a splitting point \(\overline{b}^{(t)}_{f}\), we first compute the following values \(G^{(1)}_{f,j}\), \(G^{(2)}_{f,j}\), \(G^{(3)}_{f,j}\) and \(G^{(4)}_{f,j}\) by iterating over the bins instead of over the instances, with \(f=1\ldots F\) and \(j=1,\ldots , T_f\):
$$\begin{aligned} G^{(1)}_{f,j}= & {} \sum _{b^{(t)}_{f} \in BB_f} \mu _{A_{f,j}}(\overline{b}^{(t)}_{f}) \cdot \varOmega _3^{t,f} \end{aligned}$$
$$\begin{aligned} G^{(2)}_{f,j}= & {} \sum _{b^{(t)}_{f} \in BB_f} \mu ^2_{A_{f,j}}(\overline{b}^{(t)}_{f}) \cdot \varOmega _3^{t,f} \end{aligned}$$
$$\begin{aligned} G^{(3)}_{f,j}= & {} \sum _{b^{(t)}_{f} \in BB_f} \mu _{A_{f,j}}(\overline{b}^{(t)}_{f}) \cdot \varOmega _1^{t,f} \end{aligned}$$
$$\begin{aligned} G^{(4)}_{f,j}= & {} \sum _{b^{(t)}_{f} \in BB_f} \mu ^2_{A_{f,j}}(\overline{b}^{(t)}_{f}) \cdot \varOmega _2^{t,f} \end{aligned}$$
Using \(G^{(1)}_{f,j},\) \(G^{(2)}_{f,j},\) \(G^{(3)}_{f,j}\) and \(G^{(4)}_{f,j}\), \(FM({P_{I_f}})\) can be easily computed as:
$$\begin{aligned} FM({P_{I_f}}) = \frac{ \sum ^{K_{P_{I_f}}}_{j=1} G^{(3)}_{f,j}}{ \sum ^{K_{P_{I_f}}}_{j=1} G^{(1)}_{f,j} } \end{aligned}$$
\(\text{ FVar }(P_{I_f})\) can be obtained as:
$$\begin{aligned}FVar(P_{I_f}) &=\frac{\sum ^{K_{P_{I_f}}}_{j=1} \sum _{(\overline{b}^{(t)}_{f},y^{(t)}) \in S_{f,j}} (y^{(t)} - FM({P_{I_f}}))^2 \cdot \mu ^2_{B_{f,j}}(\overline{b}^{(t)}_{f}) }{ \sum ^{K_{P_{I_f}}}_{j=1} \sum _{(\overline{b}^{(t)}_{f},y^{(t)}) \in S_{f,j}} \mu ^2_{B_{f,j}}(\overline{b}^{(t)}_{f}) } \\ & =\frac{ \sum ^{K_{P_{I_f}}}_{j=1} G^{(4)}_{f,j} +G^{(1)}_{f,j} \cdot FM({P_{I_f}})^2 - 2 \cdot G^{(3)}_{f,j} \cdot FM({P_{I_f}})}{ \sum ^{K_{P_{I_f}}}_{j=1} G^{(2)}_{f,j}} \end{aligned}$$
Finally, \(\text{ PFGain }(P_{I_f})\) is computed with Eq. 4. The output of PFV is a fuzzy partition for each input variable. These fuzzy partitions are input to the DFRT learning process.

In DFRT learning, the best split for each node is computed in parallel across the CUs. Figure 4 describes the workflow of the DFRT learning. The DFRT learning is an iterative process, which executes a complete Map-Reduce step at each iteration.

Let H be the number of iterations performed by the algorithm and h be the index of the h-th iteration. Let R be the list of nodes to be split, initialized with only one element consisting of the root of the tree. The algorithm iteratively determines the group \(R_h\) of \(\vert R \vert\) nodes from R. Finally, it performs a Map-Reduce step for distributing the growing process of the tree.

Each Map task, using the corresponding chunk v of data, computes, for each node \(NT_y \in R_h\), a vector \(D_{v,y}\) of size \(\vert D \vert\) = \(\sum _{\forall f \in F} T_f\) elements. In particular, each element j of \(D_{v,y}\) corresponds to the fuzzy set \(A_{f,j}\) and its value is equal to \(D_{v,y}[j]=\sum _{i=1}^{\left| S_{f,j}\right| } TN(\mu _{A_{f,j}}(x_{f,i})\cdot \mu _{NT_y}(\mathbf x _{i}))\), where \(\mu _{NT_y}(\mathbf x _{i})\) is the membership degree of instance \(\mathbf x _{i}\) to the parent node (for the root of the decision tree, the membership value is equal to 1) and the operator TN is a T-norm. The output of these Map tasks are key-value pairs \(\langle key = y, value = D_{v,y} \rangle\) (where y is the index of the node \(NT_y\)). Afterward, each Reduce task receives a list of vectors \(D_{v,y}\) and creates a vector \(D_y\) by performing an element-wise addition of all V vectors. Thus, \(D_y\) stores the cardinality of each input variable from the root to \(NT_y\) along the overall training set. Then, the Reduce task generates two child nodes and tests the stopping conditions described in “FRT learning” section for both of them. The children generated from each \(NT_y\) are used to update the tree and R: if a child node is not labeled as leaf, then it is inserted into the list R and employed at the next iterations. The algorithm repeats all the steps until R is empty.

In order to compute the PFGain(\(P_{I_f}\)) efficiently (used to test the second stopping criteria), we first compute the values \(G^{(1)}_{f,j}\), \(G^{(2)}_{f,j}\), \(G^{(3)}_{f,j}\) and \(G^{(4)}_{f,j}\) (Eqs. 69 respectively), similar to the distributed PFV algorithm. Further, \(\mathbf {\varOmega ^{l,f}}\) is already available because it was computed by the distributed PFV algorithm and \(\mu _{B_{f,j}}(\overline{b}^{(t)}_{f})\) is obtained efficiently by computing the T-Norm using the membership degrees stored in \(D_y\).
Fig. 4

Distributed fuzzy regression tree learning workflow

Results and discussion

In order to test the behavior of the proposed DFRT, we have performed a set of experiments that evaluate the performance in terms of MSE, complexity and scalability related to the size of the datasets and the number of used CUs. MSE is computed as:
$$\begin{aligned} \text{ MSE } = \frac{\sum _{i=1}^{\vert DS \vert } (y_i - \text{ DFRT }(\mathbf {x}_i))^2}{\vert DS \vert } \end{aligned}$$
where DS stands for each dataset.

Complexity is calculated as the number of nodes of the tree. Scalability is presented by plotting the average execution times versus the number of CUs.

We have used 8 real world datasets [2].2 These datasets come from the same real-world problem, the Protein Structure Prediction (PSP), which aims to predict the 3D structure of a protein (output variable) based on amino-acid structural continuous variables (inputs). As described in [1], a protein structure (PS) is the three-dimensional arrangement of atoms in a protein molecule. These structures arise because particular sequences of amino-acids in polypeptide chains fold to generate, from linear chains, compact domains with specific 3D structures. The folded domains can serve as modules for building up large assemblies such as virus particles or muscle fibers, or they can provide specific catalytic or binding sites, as found in enzymes or proteins that carry oxygen or regulate the function of DNA. PSP predicts the three-dimensional structures of a protein by using its first structure, its amino-acid sequence, to predict its folding and its secondary, tertiary and quaternary structure [44, 51]. This makes PSP an essential tool in proteomics since the molecular function of a protein depends on its threedimensional structure, which is often unknown.

The difference among these datasets relies on the number of input variables, varying from 100 (Window2) to 380 (Window9) by steps of 40 input variables. The number of instances is 257560 for all the datasets. To validate the results we have used a ten fold cross-validation (10-CV), being 234638 instances for training and 22922 for test at each iteration.

To the best of our knowledge few real-world big datasets exist for regression problems. Thus, with the aim of enriching the experimental evaluation and extracting more robust conclusions, we have generated 4 synthetic datasets, following the procedure used in [40]: two of them (\(f_1\) and \(f_{19}\), described in Eqs. 13 and 14) are characterized by 1000 input variables and \(10^5\) instances, and the remaining two (\(f_2\) and \(f_4\), described in Eqs. 15 and 16) by 100 input variables and \(10^6\) instances. The generation process consists of generating random instances, where each input variable has a predefined domain, and applying a function over the selected input variables to produce the output value. Let \(\mathbf {x} = \{x_1, x_2, \ldots , x_F\}\) be a random input instance in a predefined domain, \(\mathbf {P}\) a random permutation of {1, 2, …, F} and M an orthogonal matrix of dimension \(m \times m\), where m has been fixed to 50. The definitions of \(f_1\), \(f_{19}\), \(f_2\) and \(f_4\) are, respetively:
$$\begin{aligned}&f_1(\mathbf {x}) = \sum _{i=1}^F (10^6)^{\frac{i-1}{F-1}} \cdot x_i^2 \end{aligned}$$
$$\begin{aligned}&\quad f_{19}(\mathbf {x}) = \sum _{i=1}^F \left( \sum _{j=1}^i x_i \right) ^2 \end{aligned}$$
$$\begin{aligned}&\quad f_2(\mathbf {x}) = \sum _{i=1}^F \left[ x_i^2 - 10 \cos (2 \pi x_i) + 10 \right] \end{aligned}$$
$$\begin{aligned}f_4(\mathbf {x}) & = f_1( \{x_{P_1}, \ldots , x_{P_m}\} \times M ) \cdot 10^6 \\&\quad + f_1(\{x_{P_{m+1}}, \ldots , x_{P_F})\} \end{aligned}$$
To give a glimpse of the complexity of the problems, we have computed three metrics proposed in [28] for evaluating the correlation between input variables and output. In particular, we have adopted Maximum and Average Feature Correlation to the Output, and Collective Feature Efficiency, denoted as \(C_1\), \(C_2\) and \(C_4\), respectively, in the original paper. \(C_1\) and \(C_2\) measure the maximum and average Spearman correlation values between the input variables and the output (see Eqs. 17 and 18). \(C_4\) measures the percentage of instances which are difficult to predict (large residual values) with respect to a linear regressor (see Eq. 19).
$$\begin{aligned} C_1 = \max _{j=1, \ldots , F} \vert \rho (x_j, y) \vert \end{aligned}$$
$$\begin{aligned} C_2 = \sum _{j=1}^{F} \frac{\vert \rho (x_j, y) \vert }{F} \end{aligned}$$
$$\begin{aligned} C_4 = 1 - \sum _{l} \frac{s_l}{\vert DS \vert } \end{aligned}$$
where \((x_j,y)\) is the projection of the dataset DS along the input variable \(X_j\), \(\rho (x_j, y)\) is the Spearman correlation between the input variable \(X_j\) and the output, and \(s_l\) is the number of instances which have been iteratively marked as linearly classifiable (residual obtained by a linear classifier between the feature \(X_j\) and Y smaller than 0.01), being \(X_j\) the most correlated feature with respect to Y that has not been yet selected.
The properties for each dataset (number F of input variables, number \(\vert DS \vert\) of instances, and metrics \(C_1\), \(C_2\) and \(C_4\)) are shown in Table 1. The symbols \(\uparrow\) and \(\downarrow\) denote that, respectively, high (direct relation) and low (indirect relation) values of the metrics correspond to high complexity. The values highlight that the problems are particularly difficult since there does not exist a high correlation between input variables and output. In particular, the values of \(C_4\) point out that the output is very difficult to predict by using a linear regressor. In fact, we made some additional experimentation using the distributed linear least squares approach implemented in the Apache Spark MLlib. The results achieved by this algorithm were very poor and thus, for the sake of brevity, we do not present them in this work.
Table 1

Properties of the real-world and synthetic datasets



\(\vert DS \vert\)

\(C_1 (\downarrow )\)

\(C_2 (\downarrow )\)

\(C_4 (\uparrow )\)









































































The experiments have been executed on a cluster of computers with one master and six slave nodes equipped with dual Intel Xeon E5-2609v3 1.90 GHz hexacore processors and 64GB of RAM per node. Each worker node is running the HDFS file system on 4x1TB disks and is managed by the 1.5.2 version of the Apache Spark standalone distribution.

In order to validate the results of our DFRT, we have first compared them to the ones obtained by the distributed regression tree (DRT) of the MLlib Apache Spark’s implementation (from version 1.5.2). The parameters used for the execution of DRT and DFRT are shown in Table 2. The comparison is performed in terms of MSE, execution time and scalability. Then, we have compared the results of our DFRT with the ones achieved by the Distributed Random Forests (DRFs) of the MLlib Apache Spark’s implementation. Random Forest [4] is an ensemble of regression trees. The effect of using an ensemble is that, asymptotically, the bias of the model is similar to the one obtained by one single tree, while the variance is significantly reduced. As this is one of the key points in fuzzy regression trees, we have compared our DFRT to DRF in terms of variance, in addition to MSE. In “Comparison with Distributed Random Forest” section, we present this comparison and draw some conclusions.
Table 2

Parameters for the MLlib DRT and the DFRT




Fuzzy discretization

\(\lambda _{\text{disc}}\)


\(\epsilon _{\text{disc}}\)


\(\gamma _{\text{disc}}\)


Regression tree






L (maximum number of bins)



\(\lambda _{\text{learn}}\)



\(\epsilon _{\text{learn}}\)



Comparison with DRT

We will analyze the results for the real and the synthetic datasets separately (first the real-world problems). Tables 3 and 5 show the average training and test MSEs obtained by our DFRT and the MLlib DRT.
Table 3

Average MSEs obtained by DFRT and DRT on the training and test sets for the real-world problems
















































The best results for each dataset have been emphasized in italic font

As we can observe in Table 3, DRT suffers from overfitting. Indeed, the training MSE is much lower than the test MSE. Further, the test MSE obtained by DRT is much higher than the one achieved by DFRT. On the other hand, the training and test MSEs obtained by DFRT are similar, thus showing that DFRT is not prone to overfitting.

Regarding the complexity, Table 4 shows the average number of nodes (inner and leaves) and the average number of leaves. We can observe that the complexity of the models is very similar for the same algorithm for all the real-world datasets (we just observe a slight increase on the numbers of nodes and leaves in the case of DRT). From Table 1 we can observe that the real-world datasets have the same value of the \(C_1\) metric and the \(C_2\) metric decreases with the increase of the number of inputs (from Window\(_2\) to Window\(_9\)). This means that the most correlated input variables belong to the shared 100 input variables (variables from Window\(_2\)). For this reason, the results do not improve notably by including the rest of the input variables, and therefore are very similar for these datasets. Comparing the algorithms, DRT is much more complex than DFRT in terms of numbers of both nodes and leaves (approximately a difference of one order of magnitude).
Table 4

Average complexity of the models obtained by DFRT and DRT for the real-world problems
















































The best results for each dataset have been emphasized in italic font

In relation to the synthetic datasets, Table 5 shows the results grouped in two blocks: \(f_1\) and \(f_{19}\) (problems with 1000 input variables and \(10^5\) instances) and \(f_2\) and \(f_4\) (problems with 100 input variables and \(10^6\) instances). We can observe that DFRT outperforms DRT in both training and test sets. Further, we note that DRT suffers from overfitting in \(f_4\) and more evidently in \(f_{19}\). On the other side, DFRT seems not to be affected by this type of problem, thanks to the use of fuzzy rather than crisp boundaries of the decision subspaces. Table 6 shows the average numbers of nodes (again inner and leaves) and leaves. The difference of complexity is more remarkable in this case, being DRT about two orders of magnitude more complex than DFRT.
Table 5

Average MSEs obtained by DFRT and DRT on the training and test sets for the synthetic datasets




























The best results for each dataset have been emphasized in italic font

The results have to be multiplied by \(10^{22}\), \(10^{22}\), \(10^5\) and \(10^{33}\) for \(f_1\), \(f_{19}\), \(f_2\) and \(f_4\) respectively

Table 6

Average complexity of the models obtained by DFRT and DRT for the synthetic datasets




























The best results for each dataset have been emphasized in italic font

The analysis of Tables 4 and 6 highlights the capability of the DFRT learning approach to deal with high dimensionality. Indeed, in the DFRTs generated by the learning, a relevant number of input variables are missing, thus proving the input variable selection implicitly performed during the regression tree learning. If, for instance we analyze the synthetic datasets generated from \(f_1\) and \(f_{19}\), we can realize that we have 1000 input variables and, respectively, 634.9 and 678.8 nodes on average. Since the number of nodes also includes the average number of leaves, respectively 317.9 and 339.9 for the two datasets, we have that on average the two DFRTs contain 317 and 338.9 decision nodes. This implies that at most (if a different variable is used in each decision node) 317 and 338.9 out of 1000 input variables are used on average in the DFRT. Thus, we can conclude that a relevant input variable selection is performed, since the number of nodes is lower than the number of input variables. As it occurs for all the regression decision trees, we can conclude that our approach can cope with a high number of input variables by reducing this number during the learning phase.

As regards scalability, we evaluate it with respect to the number of CUs and the dimensionality of the datasets. Figures 5 and 6 show the average execution times (in seconds) of DFRT (red lines) and DRT (black lines) for the real world and synthetic datasets, respectively, versus the number of CUs. In the figures, we also show the ideal behavior of the two algorithms (dashed line), whose speedup is linear with respect to the number of CUs.
Fig. 5

Average execution times of the DRT and DFRT algorithms for the real-world datasets versus the number of CUs.

Fig. 6

Average execution times of the DRT and DFRT algorithms for the synthetic datasets versus the number of CUs.

As we can observe, in the real datasets, DFRT is in general more time consuming, specially when using a small number of CUs. However, as the number of CUs increases, both algorithms need similar computation times.

In the synthetic datasets, we can draw similar conclusions. Furthermore, we notice that, in the case of datasets with a lower number of variables, DRT is much faster than DFRT, specially when using a small number of CUs. In the case of the datasets with a higher number of variables, computation times are similar and, for some number of CUs, DFRT is faster.

There are mainly two aspects which affect the computation time. First, since the tree learning stage consists of an iterative process, the larger the number of nodes of the model, the higher the computation time. Second, the map-reduce phase searches for the best split at each iteration. This is done by analyzing all the possible splits among the candidate nodes. Hence, if the generated model is more complex, a larger number of nodes will be evaluated at each iteration with corresponding increase of the computation time. For high-dimensional datasets, DRT tends to build more complex models (a larger number of nodes) compared to DFRT, thus increasing the computation time. However, for each candidate split point, the computation of the PFGain in DFRT is much more complex than the computation of the corresponding measure in DRT and this increases drastically the computation time of DFRT with respect to DRT. Thus, in the case of datasets with a small number of input variables, DRT is generally faster than DFRT. On the other hand, in the case of datasets with a larger number of input variables, computation times are more similar to each other and sometimes DFRT can outperform DRT (for instance, in the \(f_1\) dataset).

With respect to the number of CUs, we expect that the larger the number of CUs, the lower the computation time. However, this behavior is affected by the overhead due to communication. We can see how the ratio ideal versus real speedup is directly dependent on the number of CUs: the larger the number of workers, the worse the ratio. This happens because the increase of CUs requires a larger communication among workers and therefore a higher overhead, which is easily observed in the case of the synthetic datasets, where an increase in the number of CUs upon 18 requires a higher computational time than using a lower number of CUs.

However, this behavior is not observed in the case of real-world problems. The reason is that the synthetic datasets have been generated uniformly over the input space, unlike the real-world problems. We observed that, in this scenario, the first phase of the distributed PFV, namely the Distributed Bin Generation, tends to produce a sorted list of bin boundaries per each variable with the maximum number of elements, namely \(V \cdot L\). This implies that the second Map-Reduce phase, namely the Distributed Fuzzy Sets Generation, will require a lot of communication to transfer the list of bin boundaries to any worker node for computing the statistics during the execution of the Map task and then to collect the statistics during the execution of the Reduce task. With the increase of the number of nodes, the overhead due to this communication grows and makes the advantage of having a higher number of CUs ineffective.

Comparison with Distributed Random Forest

In this section, we aim to compare the results obtained by DFRT with one of the state-of-the-art algorithms available for regression problems with big data, namely DRF. We adopt DRF for two reasons. First, DRF is available in the Spark MLlib library. Second, it has been proved that, comparing an RT versus an RF (ensemble of RTs) [41], the bias of RF is asymptotically similar but the variance is reduced. Thus, we will compare DFRT with DRF with respect to variance, in addition to MSE (we will use the standard deviation to improve the readability of the results).

We execute DRF by using the default values recommended by the authors. Furthermore, we have set the two parameters, which do not have default values, namely the maximum depth for any tree and the number of trees, which compose the ensemble, to 10 (parameter \(\beta\)) and 20, respectively. In a preliminary experimentation, we tried different (and larger) values for these parameters. However, we could not run the algorithm on a single CU because of memory overflow. In order to compare the algorithms under similar environments, we finally chose the previous parametrization.

Table 7 shows the average training and test MSEs with their respective standard deviations. We can observe that DFRT achieves a lower test MSE than DRF in the real world problems, while DRF obtains a lower test MSE than DFRT in the synthetic datasets. In the real datasets, we observe that RF is affected by overfitting (the test MSE is larger than almost 30% of the training MSE). However, in the case of synthectic datasets, training and test MSEs are very similar for DFRT and RF. We have applied a Wilcoxon Signed Rank Test [46] to evaluate whether there exists a statistical difference between the two algorithms. We adopted \(\alpha = 0.05\) as confidence level and obtained p-value=0.1697, which points out that there is no statistical evidence that these algorithms are different.

As regards the standard deviation, DFRTs are characterized by smaller standard deviations than DRFs, except for \(f_{19}\). Moreover, the standard deviation obtained by DFRT is small (always less than a 10% of the MSE), while in the case of RF it is greater, up to 20% in the case of real world problems. Again, we have applied a Wilcoxon Signed Rank Test with \(\alpha = 0.05\). We have obtained \(\text{ p-value } = 0.0046\), thus pointing out that the standard deviation obtained by DFRT is statistically lower than the one achieved by DRFs. This result is very interesting, considering the observation made in [41] regarding the variance of the MSEs obtained by RFs.
Table 7

Average MSEs and their standard deviation obtained by DFRT and RF on the training and test sets for the real world and synthetic datasets




Training ± s.d.

Test ± s.d.

Training ± s.d.

Test ± s.d.


24.3263 ± 0.18

24.3978 ± 0.90

22.9981 ± 3.02

29.9958 ± 5.72


24.2976 ± 0.18

24.3853 ± 0.91

23.1057 ± 3.43

29.0574 ± 5.35


24.2841 ± 0.18

24.3730 ± 0.91

21.3935 ± 1.31

28.9095 ± 5.13


24.2833 ± 0.18

24.3732 ± 0.91

22.1959 ± 2.85

28.8784 ± 4.95


24.2834 ± 0.18

24.3733 ± 0.91

22.5899 ± 3.53

29.0761 ± 5.52


24.2836 ± 0.18

24.3738 ± 0.91

21.9378 ± 2.56

29.7301 ± 4.78


24.2841 ± 0.18

24.3742 ± 0.91

21.0925 ± 2.64

31.1765 ± 5.26


24.2807 ± 0.18

24.3689 ± 0.91

21.8789 ± 2.82

28.9022 ± 4.94


15.0433 ± 0.03

16.6537 ± 0.18

17.6090 ± 0.48

15.0783 ± 0.84


6.5671 ± 0.02

7.3154 ± 0.11

6.6059 ± 0.08

6.8696 ± 0.09


31.3784 ± 0.02

32.0971 ± 0.14

32.2631 ± 0.28

29.1347 ± 0.75


428.8752 ± 0.47

429.8814 ± 3.39

426.1867 ± 0.74

427.2893 ± 3.40

The best results for each dataset have been emphasized in italic font

The results have to be multiplied by \(10^{21}\), \(10^{22}\), \(10^4\) and \(10^{31}\) for \(f_1\), \(f_{19}\), \(f_2\) and \(f_4\) respectively


In this work we have proposed a distributed fuzzy regression tree learning algorithm (DFRT), following the Map Reduce programming model to cope with Big Data problems. DFRT uses a novel distributed fuzzy discretizer (PFV), adapted from [38] to deal with regression problems. PFV generates a strong fuzzy partition based on the fuzzy variance. Then, DFRT applies an iterative process to generate a binary fuzzy regression tree. The distributed version of the algorithm has been implemented using the Apache Spark framework.

We have used 8 real-world datasets, generated from the same problem. In particular, all of them have the same number of instances, but they differ in the number of input variables. In addition, we have generated 4 synthetic datasets: two of them with 1000 input variables and \(10^5\) instances, and the others with 100 input variables and \(10^6\) instances. We have compared the results obtained by our approach with the distributed regression tree (DRT) and the Distributed Random Forests (DRFs) implemented in the MLlib Apache Library (included in the 1.5.2 version of the Apache Spark).

Results show that DFRT scales similar to the DRT algorithm, but the results obtained by DFRT outperform in all the cases the DRT in terms of MSE on the test set and also in terms of complexity (number of nodes and leaves). In terms of execution time, even if the scalability behavior is similar to DRT, the DFRT algorithm is slower. We have observed that, in the case of the smaller datasets (less number of features), DRT is generally faster than DFRT. On the other hand, DFRT is faster (or at least similar) than DRT for the case of datasets with a large number of input variables.

As regards the comparison with DRF, the results show that DFRT and DRF are not statistically different in terms of test MSE, but DFRT outperforms DRF in terms of standard deviation.


  1. 1.

    In this work, we use the product as T-Norm.

  2. 2.


Authors’ contributions

Authors propose a distributed fuzzy regression tree learning algorithm following the Map Reduce programming model to cope with Big Data problems. Authors adapt a novel distributed fuzzy discretizer (PFV) from [38] to deal with regression problems. The algorithm has been implemented using the 1.5.2 version of Apache Spark framework. All authors read and approved the final manuscript.


The authors would like to acknowledge Dr. Armando Segatori for the precious suggestions and explanations on the implementation of the distributed fuzzy decision trees for big data classification problems.

Competing interests

The authors declare that they have no competing interests.

Availability of data and materials

Real world datasets can be downloaded from the website To download the synthetic datasets, please contact the corresponding author Javier Cózar


This work has been partially supported by the project PRA 2017 “IoT e Big Data: metodologie e tecnologie per la raccolta e l’elaborazione di grosse moli di dati”, funded by the University of Pisa. This work has been partially funded by the Spanish Research Agency (AEI/MINECO) and FEDER (UE) through projects TIN2013-46638-C3-3-P and TIN2016-77902-C3-1-P. This work has been partially funded by the Junta de Comunidades de Castilla-La Mancha and FEDER (UE) funds through project SBPLY/17/180501/000493. Javier Cózar has also been funded by the MICINN grant FPU12/05102.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.


  1. 1.
    Arana-Daniel N, Gallegos AA, López-Franco C, Alanís AY, Morales J, López-Franco A. Support vector machines trained with evolutionary algorithms employing kernel adatron for large scale classification of protein structures. Evol Bioinform. 2016;12:285–302.CrossRefGoogle Scholar
  2. 2.
    Bacardit J, Krasnogor N. The icos psp benchmarks repository; 2008. Accessed 3 Dec 2018.
  3. 3.
    Berzal F, Cubero JC, Marın N, Sánchez D. Building multi-way decision trees with numerical attributes. Inf Sci. 2004;165(1):73–90.MathSciNetzbMATHCrossRefGoogle Scholar
  4. 4.
    Breiman L. Random forests. Mach Learn. 2001;45(1):5–32.zbMATHCrossRefGoogle Scholar
  5. 5.
    Breiman L, Friedman JH, Olshen RA, Stone CJ. Classification and regression trees. Monterey: Wadsworth & Brooks; 1984.zbMATHGoogle Scholar
  6. 6.
    Cheng HD, Chen JR. Automatically determine the membership function based on the maximum entropy principle. Inf Sci. 1997;96(3–4):163–82.CrossRefGoogle Scholar
  7. 7.
    Cózar J, delaOssa L, Gámez JA. Learning tsk-0 linguistic fuzzy rules by means of local search algorithms. Appl Soft Comput. 2014;21:57–71.CrossRefGoogle Scholar
  8. 8.
    Cózar J, delaOssa L, Gámez JA. Tsk-0 fuzzy rule-based systems for high-dimensional problems using the apriori principle for rule generation. In: Rough sets and current trends in computing, lecture notes in computer science, vol 8536. New York: Springer International Publishing; 2014. p. 270–9.Google Scholar
  9. 9.
    Dean J, Ghemawat S. Mapreduce: simplified data processing on large clusters. Commun ACM. 2008;51(1):107–13.CrossRefGoogle Scholar
  10. 10.
    Slowiński R. Fuzzy sets in decision analysis, operations research and statistics, vol 1. US: Springer; 2012.Google Scholar
  11. 11.
    Diao R, Sun K, Vittal V, O’Keefe RJ, Richardson MR, Bhatt N, Stradford D, Sarawgi SK. Decision tree-based online voltage security assessment using pmu measurements. IEEE Trans Power Syst. 2009;24(2):832–9.CrossRefGoogle Scholar
  12. 12.
    Dubois D, Prade H. Unfair coins and necessity measures: towards a possibilistic interpretation of histograms. Fuzzy Sets and Systems. 1983;10(1–3):15–20.MathSciNetzbMATHCrossRefGoogle Scholar
  13. 13.
    Fonarow GC, Adams KF, Abraham WT, Yancy CW, Boscardin WJ, Committee ASA. Risk stratification for in-hospital mortality in acutely decompensated heart failure: classification and regression tree analysis. JAMA. 2005;293(5):572–80.CrossRefGoogle Scholar
  14. 14.
    Franklin J. The elements of statistical learning: data mining, inference and prediction. Math Intell. 2005;27(2):83–5.CrossRefGoogle Scholar
  15. 15.
    Garcia S, Luengo J, Sáez JA, Lopez V, Herrera F. A survey of discretization techniques: taxonomy and empirical analysis in supervised learning. IEEE Trans Knowl Data Eng. 2013;25(4):734–50.CrossRefGoogle Scholar
  16. 16.
    Goetz T. The decision tree: taking control of your health in the new era of personalized medicine. Emmaus: Rodale; 2010.Google Scholar
  17. 17.
    Gupta A, Mehrotra KG, Mohan C. A clustering-based discretization for supervised learning. Stat Probab Lett. 2010;80(9):816–24.MathSciNetzbMATHCrossRefGoogle Scholar
  18. 18.
    Haskell RE. Regression tree fuzzy systems. In: Proceedings of the ICSC symposium on soft computing, fuzzy logic, artificial neural networks and genetic algorithms, University of Reading, Whiteknights, Reading, England; 1996. p. 26–8.Google Scholar
  19. 19.
    Hastie T, Tibshirani R, Friedman J. The elements of statistical learning: data mining, inference, and prediction. Springer series in statistics, vol 10, 1st ed. New York: Springer; 2001.zbMATHCrossRefGoogle Scholar
  20. 20.
    Izrailev S, Agrafiotis D. A novel method for building regression tree models for qsar based on artificial ant colony systems. J Chem Inf Comput Sci. 2001;41(1):176–80.CrossRefGoogle Scholar
  21. 21.
    Keller JM, Gray MR, Givens JA. A fuzzy k-nearest neighbor algorithm. IEEE Trans Syst Man Cybern. 1985;4:580–5.CrossRefGoogle Scholar
  22. 22.
    Kim H, Loh WY. Classification trees with unbiased multiway splits. J Am Stat Assoc. 2001;96(454):589–604.MathSciNetCrossRefGoogle Scholar
  23. 23.
    Kotsiantis S, Kanellopoulos D. Discretization techniques: a recent survey. GESTS Int Trans Comput Sci Eng. 2006;32(1):47–58.Google Scholar
  24. 24.
    Leathwick J, Elith J, Francis M, Hastie T, Taylor P. Variation in demersal fish species richness in the oceans surrounding new zealand: an analysis using boosted regression trees. Mar Ecol Prog Ser. 2006;321:267–81.CrossRefGoogle Scholar
  25. 25.
    Leskovec J, Rajaraman A, Ullman JD. Mining of massive datasets. Cambridge: Cambridge University Press; 2014.CrossRefGoogle Scholar
  26. 26.
    Leskovec J, Rajaraman A, Ullman JD. Mining of massive datasets. Cambridge: Cambridge university press; 2014.CrossRefGoogle Scholar
  27. 27.
    Liu H, Hussain F, Tan CL, Dash M. Discretization: an enabling technique. Data Mining Knowl Discov. 2002;6(4):393–423.MathSciNetCrossRefGoogle Scholar
  28. 28.
    Maciel AI, Costa IG, Lorena AC Measuring the complexity of regression problems. In: 2016 international joint conference on neural networks (IJCNN). New York: IEEE; 2016. p. 1450–7.Google Scholar
  29. 29.
    Medasani S, Kim J, Krishnapuram R. An overview of membership function generation techniques for pattern recognition. Int J Approx Reason. 1998;19(3–4):391–417.MathSciNetzbMATHCrossRefGoogle Scholar
  30. 30.
    Meng X. Mllib: Scalable machine learning on spark. In: Spark Workshop April; 2014.Google Scholar
  31. 31.
    Mori H, Kosemura N, Ishiguro K, Kondo T. Short-term load forecasting with fuzzy regression tree in power systems. In: 2001 IEEE international conference on systems, man, and cybernetics, vol 3. New York: IEEE; 2001. p. 1948–53.Google Scholar
  32. 32.
    Nieradka G, Butkiewicz B. A method for automatic membership function estimation based on fuzzy measures. In: International fuzzy systems association world congress. Berlin: Springer; 2007. p. 451–60Google Scholar
  33. 33.
    Olaru C, Wehenkel L. A complete fuzzy decision tree technique. Fuzzy Sets Syst. 2003;138(2):221–54.MathSciNetCrossRefGoogle Scholar
  34. 34.
    Pedrycz W. Why triangular membership functions? Fuzzy Sets Syst. 1994;64(1):21–30.MathSciNetCrossRefGoogle Scholar
  35. 35.
    Prasad AM, Iverson LR, Liaw A. Newer classification and regression tree techniques: bagging and random forests for ecological prediction. Ecosystems. 2006;9(2):181–99.CrossRefGoogle Scholar
  36. 36.
    Quinlan RJ. Learning with continuous classes. In: 5th Australian joint conference on artificial intelligence. Singapore: World Scientific; 1992. p. 343–8.Google Scholar
  37. 37.
    Segal MR. Regression trees for censored data. Biometrics. 1988;44:35–47.zbMATHCrossRefGoogle Scholar
  38. 38.
    Segatori A, Marcelloni F, Pedrycz W. On distributed fuzzy decision trees for big data. IEEE Trans Fuzzy Syst. 2018;26(1):174–92.CrossRefGoogle Scholar
  39. 39.
    Suárez A, Lutsko JF. Globally optimal fuzzy decision trees for classification and regression. IEEE Trans Pattern Anal Mach Intell. 1999;21(12):1297–311.CrossRefGoogle Scholar
  40. 40.
    Tang K, Li X, Suganthan PN, Yang Z, Weise T. Benchmark functions for the cec2010 special session and competition on large-scale global optimization. Tech. rep. nature inspired computation and applications laboratory; 2009.Google Scholar
  41. 41.
    Wager S Asymptotic theory for random forests. arXiv preprint; 2014. arXiv:14050352.
  42. 42.
    Ward JS, Barker A. Undefined by data: a survey of big data definitions. arXiv preprint; 2013. arXiv:13095821.
  43. 43.
    Weber R. Fuzzy-id3: a class of methods for automatic knowledge acquisition. In: Proceedings of the 2nd international conference on fuzzy logic and neural networks; 1992.Google Scholar
  44. 44.
    Westhead DR, Thornton JM. Protein structure prediction. Curr Opin Biotechnol. 1998;9(4):383–9.CrossRefGoogle Scholar
  45. 45.
    White T. Hadoop: the definitive guide. Sebastopol: O’Reilly Media, Inc.; 2012.Google Scholar
  46. 46.
    Wilcoxon F. Individual comparisons by ranking methods. Biometrics Bull. 1945;1(6):80–3.CrossRefGoogle Scholar
  47. 47.
    Yuan Y, Shaw MJ. Induction of fuzzy decision trees. Fuzzy Sets Syst. 1995;69(2):125–39.MathSciNetCrossRefGoogle Scholar
  48. 48.
    Zaharia M. Apache Spark MLlib; 2009. Accessed 26 Sept 2017.
  49. 49.
    Zaharia M, Chowdhury M, Franklin MJ, Shenker S, Stoica I. Spark: cluster computing with working sets. HotCloud. 2010;10(10–10):95.Google Scholar
  50. 50.
    Zeinalkhani M, Eftekhari M. Fuzzy partitioning of continuous attributes through discretization methods to construct fuzzy decision tree classifiers. Inf Sci. 2014;278:715–35.MathSciNetzbMATHCrossRefGoogle Scholar
  51. 51.
    Zhang Y, Skolnick J. The protein structure prediction problem could be solved using the current pdb library. Proc Natl Acad Sci. 2005;102(4):1029–34.CrossRefGoogle Scholar
  52. 52.
    Zheng Y, Liu L, Wang L, Xie X. Learning transportation mode from raw gps data for geographic applications on the web. In: Proceedings of the 17th international conference on World Wide Web, New York: ACM; 2008. p. 247–56.Google Scholar

Copyright information

© The Author(s) 2018

Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (, which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

Authors and Affiliations

  1. 1.Departamento de Sistemas Informáticos/I3AUniversidad de Castilla-La ManchaAlbaceteSpain
  2. 2.Dipartimento di Ingegneria dell’InformazionePisaItaly

Personalised recommendations