Abstract
Time Series Classification (TSC) is an important and challenging problem in data mining. With the increase of time series data availability, hundreds of TSC algorithms have been proposed. Among these methods, only a few have considered Deep Neural Networks (DNNs) to perform this task. This is surprising as deep learning has seen very successful applications in the last years. DNNs have indeed revolutionized the field of computer vision especially with the advent of novel deeper architectures such as Residual and Convolutional Neural Networks. Apart from images, sequential data such as text and audio can also be processed with DNNs to reach stateoftheart performance for document classification and speech recognition. In this article, we study the current stateoftheart performance of deep learning algorithms for TSC by presenting an empirical study of the most recent DNN architectures for TSC. We give an overview of the most successful deep learning applications in various time series domains under a unified taxonomy of DNNs for TSC. We also provide an open source deep learning framework to the TSC community where we implemented each of the compared approaches and evaluated them on a univariate TSC benchmark (the UCR/UEA archive) and 12 multivariate time series datasets. By training 8730 deep learning models on 97 time series datasets, we propose the most exhaustive study of DNNs for TSC to date.
Introduction
During the last two decades, Time Series Classification (TSC) has been considered as one of the most challenging problems in data mining (Yang and Wu 2006; Esling and Agon 2012). With the increase of temporal data availability (Silva et al. 2018), hundreds of TSC algorithms have been proposed since 2015 (Bagnall et al. 2017). Due to their natural temporal ordering, time series data are present in almost every task that requires some sort of human cognitive process (Längkvist et al. 2014). In fact, any classification problem, using data that is registered taking into account some notion of ordering, can be cast as a TSC problem (Cristian Borges Gamboa 2017). Time series are encountered in many realworld applications ranging from electronic health records (Rajkomar et al. 2018) and human activity recognition (Nweke et al. 2018; Wang et al. 2018a) to acoustic scene classification (Nwe et al. 2017) and cybersecurity (Susto et al. 2018). In addition, the diversity of the datasets’ types in the UCR/UEA archive (Chen et al. 2015b; Bagnall et al. 2017) (the largest repository of time series datasets) shows the different applications of the TSC problem.
Given the need to accurately classify time series data, researchers have proposed hundreds of methods to solve this task (Bagnall et al. 2017). One of the most popular and traditional TSC approaches is the use of a nearest neighbor (NN) classifier coupled with a distance function (Lines and Bagnall 2015). Particularly, the Dynamic Time Warping (DTW) distance when used with a NN classifier has been shown to be a very strong baseline (Bagnall et al. 2017). Lines and Bagnall (2015) compared several distance measures and showed that there is no single distance measure that significantly outperforms DTW. They also showed that ensembling the individual NN classifiers (with different distance measures) outperforms all of the ensemble’s individual components. Hence, recent contributions have focused on developing ensembling methods that significantly outperforms the NN coupled with DTW (NNDTW) (Bagnall et al. 2016; Hills et al. 2014; Bostrom and Bagnall 2015; Lines et al. 2016; Schäfer 2015; Kate 2016; Deng et al. 2013; Baydogan et al. 2013). These approaches use either an ensemble of decision trees (random forest) (Baydogan et al. 2013; Deng et al. 2013) or an ensemble of different types of discriminant classifiers (Support Vector Machine (SVM), NN with several distances) on one or several feature spaces (Bagnall et al. 2016; Bostrom and Bagnall 2015; Schäfer 2015; Kate 2016). Most of these approaches significantly outperform the NNDTW (Bagnall et al. 2017) and share one common property, which is the data transformation phase where time series are transformed into a new feature space (for example using shapelets transform (Bostrom and Bagnall 2015) or DTW features (Kate 2016)). This notion motivated the development of an ensemble of 35 classifiers named COTE (Collective Of Transformationbased Ensembles) (Bagnall et al. 2016) that does not only ensemble different classifiers over the same transformation, but instead ensembles different classifiers over different time series representations. Lines et al. (2016, 2018) extended COTE with a Hierarchical Vote system to become HIVECOTE which has been shown to achieve a significant improvement over COTE by leveraging a new hierarchical structure with probabilistic voting, including two new classifiers and two additional representation transformation domains. HIVECOTE is currently considered the stateoftheart algorithm for time series classification (Bagnall et al. 2017) when evaluated over the 85 datasets from the UCR/UEA archive.
To achieve its high accuracy, HIVECOTE becomes hugely computationally intensive and impractical to run on a real big data mining problem (Bagnall et al. 2017). The approach requires training 37 classifiers as well as crossvalidating each hyperparameter of these algorithms, which makes the approach infeasible to train in some situations (Lucas et al. 2018). To emphasize on this infeasibility, note that one of these 37 classifiers is the Shapelet Transform (Hills et al. 2014) whose time complexity is \(O(n^2\cdot l^4)\) with n being the number of time series in the dataset and l being the length of a time series. Adding to the training time’s complexity is the high classification time of one of the 37 classifiers: the nearest neighbor which needs to scan the training set before taking a decision at test time. Therefore since the nearest neighbor constitutes an essential component of HIVECOTE, its deployment in a realtime setting is still limited if not impractical. Finally, adding to the huge runtime of HIVECOTE, the decision taken by 37 classifiers cannot be interpreted easily by domain experts, since researchers already struggle with understanding the decisions taken by an individual classifier.
After having established the current stateoftheart of non deep classifiers for TSC (Bagnall et al. 2017), we discuss the success of Deep Learning (LeCun et al. 2015) in various classification tasks which motivated the recent utilization of deep learning models for TSC (Wang et al. 2017b). Deep Convolutional Neural Networks (CNNs) have revolutionized the field of computer vision (Krizhevsky et al. 2012). For example, in 2015, CNNs were used to reach human level performance in image recognition tasks (Szegedy et al. 2015). Following the success of deep neural networks (DNNs) in computer vision, a huge amount of research proposed several DNN architectures to solve natural language processing (NLP) tasks such as machine translation (Sutskever et al. 2014; Bahdanau et al. 2015), learning word embeddings (Mikolov et al. 2013b, a) and document classification (Le and Mikolov 2014; Goldberg 2016). DNNs also had a huge impact on the speech recognition community (Hinton et al. 2012; Sainath et al. 2013). Interestingly, we should note that the intrinsic similarity between the NLP and speech recognition tasks is due to the sequential aspect of the data which is also one of the main characteristics of time series data.
In this context, this paper targets the following open questions: What is the current stateoftheart DNN for TSC? Is there a current DNN approach that reaches stateoftheart performance for TSC and is less complex than HIVECOTE? What type of DNN architectures works best for the TSC task? How does the random initialization affect the performance of deep learning classifiers? And finally: Could the blackbox effect of DNNs be avoided to provide interpretability? Given that the latter questions have not been addressed by the TSC community, it is surprising how much recent papers have neglected the possibility that TSC problems could be solved using a pure feature learning algorithm (Neamtu et al. 2018; Bagnall et al. 2017; Lines et al. 2016). In fact, a recent empirical study (Bagnall et al. 2017) evaluated 18 TSC algorithms on 85 time series datasets, none of which was a deep learning model. This shows how much the community lacks of an overview of the current performance of deep learning models for solving the TSC problem (Lines et al. 2018).
In this paper, we performed an empirical comparative study of the most recent deep learning approaches for TSC. With the rise of graphical processing units (GPUs), we show how deep architectures can be trained efficiently to learn hidden discriminative features from raw time series in an endtoend manner. Similarly to Bagnall et al. (2017), in order to have a fair comparison between the tested approaches, we developed a common framework in Python, Keras (Chollet 2015) and Tensorflow (Abadi et al. 2015) to train the deep learning models on a cluster of more than 60 GPUs.
In addition to the univariate datasets’ evaluation, we tested the approaches on 12 Multivariate Time Series (MTS) datasets (Baydogan 2015). The multivariate evaluation shows another benefit of deep learning models, which is the ability to handle the curse of dimensionality (Bellman 2010; Keogh and Mueen 2017) by leveraging different degrees of smoothness in compositional function (Poggio et al. 2017) as well as the parallel computations of the GPUs (Lu et al. 2015).
As for comparing the classifiers over multiple datasets, we followed the recommendations in Demšar (2006) and used the Friedman test (Friedman 1940) to reject the null hypothesis. Once we have established that a statistical difference exists within the classifiers’ performance, we followed the pairwise posthoc analysis recommended by Benavoli et al. (2016) where the average rank comparison is replaced by a Wilcoxon signedrank test (Wilcoxon 1945) with Holm’s alpha correction (Holm 1979; Garcia and Herrera 2008). See Sect. 5 for examples of critical difference diagrams (Demšar 2006), where a thick horizontal line shows a group of classifiers (a clique) that are not significantly different in terms of accuracy.
In this study, we have trained about 1 billion parameters across 97 univariate and multivariate time series datasets. Despite the fact that a huge number of parameters risks overfitting (Zhang et al. 2017) the relatively small train set in the UCR/UEA archive, our experiments showed that not only DNNs are able to significantly outperform the NNDTW, but are also able to achieve results that are not significantly different than COTE and HIVECOTE using a deep residual network architecture (He et al. 2016; Wang et al. 2017b). Finally, we analyze how poor random initializations can have a significant effect on a DNN’s performance.
The rest of the paper is structured as follows. In Sect. 2, we provide some background materials concerning the main types of architectures that have been proposed for TSC. In Sect. 3, the tested architectures are individually presented in details. We describe our experimental open source framework in Sect. 4. The corresponding results and the discussions are presented in Sect. 5. In Sect. 6, we describe in detail a couple of methods that mitigate the blackbox effect of the deep learning models. Finally, we present a conclusion in Sect. 7 to summarize our findings and discuss future directions.
The main contributions of this paper can be summarized as follows:

We explain with practical examples, how deep learning can be adapted to one dimensional time series data.

We propose a unified taxonomy that regroups the recent applications of DNNs for TSC in various domains under two main categories: generative and discriminative models.

We detail the architecture of nine endtoend deep learning models designed specifically for TSC.

We evaluate these models on the univariate UCR/UEA archive benchmark and 12 MTS classification datasets.

We provide the community with an open source deep learning framework for TSC in which we have implemented all nine approaches.

We investigate the use of Class Activation Map (CAM) in order to reduce DNNs’ blackbox effect and explain the different decisions taken by various models.
Background
In this section, we start by introducing the necessary definitions for ease of understanding. We then follow by an extensive theoretical background on training DNNs for the TSC task. Finally we present our proposed taxonomy of the different DNNs with examples of their application in various real world data mining problems.
Time series classification
Before introducing the different types of neural networks architectures, we go through some formal definitions for TSC.
Definition 1
A univariate time series \(X=[x_1,x_2, \dots , x_T]\) is an ordered set of real values. The length of X is equal to the number of real values T.
Definition 2
An Mdimensional MTS, \(X=[X^1,X^2, \dots , X^M]\) consists of M different univariate time series with \(X^i\)\(\in \mathbb {R}^T\).
Definition 3
A dataset \(D=\{(X_1,Y_1), (X_2,Y_2), \dots , (X_N,Y_N)\}\) is a collection of pairs \((X_i,Y_i)\) where \(X_i\) could either be a univariate or multivariate time series with \(Y_i\) as its corresponding onehot label vector. For a dataset containing K classes, the onehot label vector \(Y_i\) is a vector of length K where each element \(j\in [1,K]\) is equal to 1 if the class of \(X_i\) is j and 0 otherwise.
The task of TSC consists of training a classifier on a dataset D in order to map from the space of possible inputs to a probability distribution over the class variable values (labels).
Deep learning for time series classification
In this review, we focus on the TSC task (Bagnall et al. 2017) using DNNs which are considered complex machine learning models (LeCun et al. 2015). A general deep learning framework for TSC is depicted in Fig. 1. These networks are designed to learn hierarchical representations of the data. A deep neural network is a composition of L parametric functions referred to as layers where each layer is considered a representation of the input domain (Papernot and McDaniel 2018). One layer \(l_i\), such as \(i \in 1 \dots L\), contains neurons, which are small units that compute one element of the layer’s output. The layer \(l_i\) takes as input the output of its previous layer \(l_{i1}\) and applies a nonlinearity (such as the sigmoid function) to compute its own output. The behavior of these nonlinear transformations is controlled by a set of parameters \(\theta _i\) for each layer. In the context of DNNs, these parameters are called weights which link the input of the previous layer to the output of the current layer. Hence, given an input x, a neural network performs the following computations to predict the class:
where \(f_i\) corresponds to the nonlinearity applied at layer \(l_i\). For simplicity, we will omit the vector of parameters \(\theta \) and use f(x) instead of \(f(\theta ,x)\). This process is also referred to as feedforward propagation in the deep learning literature.
During training, the network is presented with a certain number of known inputoutput (for example a dataset D). First, the weights are initialized randomly (LeCun et al. 1998b), although a robust alternative would be to take a pretrained model on a source dataset and finetune it on the target dataset (Pan and Yang 2010). This process is known as transfer learning which we do not study empirically, rather we discuss the transferability of each model with respect to the architecture in Sect. 3. After the weight’s initialization, a forward pass through the model is applied: using the function f the output of an input x is computed. The output is a vector whose components are the estimated probabilities of x belonging to each class. The model’s prediction loss is computed using a cost function, for example the negative log likelihood. Then, using gradient descent (LeCun et al. 1998b), the weights are updated in a backward pass to propagate the error. Thus, by iteratively taking a forward pass followed by backpropagation, the model’s parameters are updated in a way that minimizes the loss on the training data.
During testing, the probabilistic classifier (the model) is tested on unseen data which is also referred to as the inference phase: a forward pass on this unseen input followed by a class prediction. The prediction corresponds to the class whose probability is maximum. To measure the performance of the model on the test data (generalization), we adopted the accuracy measure [similarly to Bagnall et al. (2017)]. One advantage of DNNs over nonprobabilistic classifiers (such as NNDTW) is that a probabilistic decision is taken by the network (Large et al. 2017), thus allowing to measure the confidence of a certain prediction given by an algorithm.
Although there exist many types of DNNs, in this review we focus on three main DNN architectures used for the TSC task: Multi Layer Perceptron (MLP), Convolutional Neural Network (CNN) and Echo State Network (ESN). These three types of architectures were chosen since they are widely adopted for endtoend deep learning (LeCun et al. 2015) models for TSC.
Multi layer perceptrons
An MLP constitutes the simplest and most traditional architecture for deep learning models. This form of architecture is also known as a fullyconnected (FC) network since the neurons in layer \(l_i\) are connected to every neuron in layer \(l_{i1}\) with \(i\in [1,L]\). These connections are modeled by the weights in a neural network. A general form of applying a nonlinearity to an input time series X can be seen in the following equation:
with \(\omega _{l_i}\) being the set of weights with length and number of dimensions identical to X’s, b the bias term and \(A_{l_i}\) the activation of the neurons in layer \(l_i\). Note that the number of neurons in a layer is considered a hyperparameter.
One impediment from adopting MLPs for time series data is that they do not exhibit any spatial invariance. In other words, each time stamp has its own weight and the temporal information is lost: meaning time series elements are treated independently from each other. For example the set of weights \(w_d\) of neuron d contains \(T\times M\) values denoting the weight of each time stamp t for each dimension of the Mdimensional input MTS of length T. Then by cascading the layers we obtain a computation graph similar to Eq. 1.
For TSC, the final layer is usually a discriminative layer that takes as input the activation of the previous layer and gives a probability distribution over the class variables in the dataset. Most deep learning approaches for TSC employ a softmax layer which corresponds to an FC layer with softmax as activation function f and a number of neurons equal to the number of classes in the dataset. Three main useful properties motivate the use of the softmax activation function: the sum of probabilities is guaranteed to be equal to 1, the function is differentiable and it is an adaptation of logistic regression to the multinomial case. The result of a softmax function can be defined as follows:
with \(\hat{Y}_j\) denoting the probability of X having the class Y equal to class j out of K classes in the dataset. The set of weights \(w_j\) (and the corresponding bias \(b_j\)) for each class j are linked to each previous activation in layer \(l_{L1}\).
The weights in Eqs. (2) and (3) should be learned automatically using an optimization algorithm that minimizes an objective cost function. In order to approximate the error of a certain given value of the weights, a differentiable cost (or loss) function that quantifies this error should be defined. The most used loss function in DNNs for the classification task is the categorical cross entropy as defined in the following equation:
with L denoting the loss or cost when classifying the input time series X. Similarly, the average loss when classifying the whole training set of D can be defined using the following equation:
with \(\varOmega \) denoting the set of weights to be learned by the network (in this case the weights w from Eqs. 2 and 3). The loss function is minimized to learn the weights in \(\varOmega \) using a gradient descent method which is defined using the following equation:
with \(\alpha \) denoting the learning rate of the optimization algorithm. By subtracting the partial derivative, the model is actually autotuning the parameters \(\omega \) in order to reach a local minimum of J in case of a nonlinear classifier (which is almost always the case for a DNN). We should note that when the partial derivative cannot be directly computed with respect to a certain parameter \(\omega \), the chain rule of derivative is employed which is in fact the main idea behind the backpropagation algorithm (LeCun et al. 1998b).
Convolutional neural networks
Since AlexNet (Krizhevsky et al. 2012) won the ImageNet competition in 2012, deep CNNs have seen a lot of successful applications in many different domains (LeCun et al. 2015) such as reaching human level performance in image recognition problems (Szegedy et al. 2015) as well as different natural language processing tasks (Sutskever et al. 2014; Bahdanau et al. 2015). Motivated by the success of these CNN architectures in these various domains, researchers have started adopting them for time series analysis (Cristian Borges Gamboa 2017).
A convolution can be seen as applying and sliding a filter over the time series. Unlike images, the filters exhibit only one dimension (time) instead of two dimensions (width and height). The filter can also be seen as a generic nonlinear transformation of a time series. Concretely, if we are convoluting (multiplying) a filter of length 3 with a univariate time series, by setting the filter values to be equal to \([\frac{1}{3},\frac{1}{3},\frac{1}{3}]\), the convolution will result in applying a moving average with a sliding window of length 3. A general form of applying the convolution for a centered time stamp t is given in the following equation:
where C denotes the result of a convolution (dot product \(*\)) applied on a univariate time series X of length T with a filter \(\omega \) of length l, a bias parameter b and a final nonlinear function f such as the Rectified Linear Unit (ReLU). The result of a convolution (one filter) on an input time series X can be considered as another univariate time series C that underwent a filtering process. Thus, applying several filters on a time series will result in a multivariate time series whose dimensions are equal to the number of filters used. An intuition behind applying several filters on an input time series would be to learn multiple discriminative features useful for the classification task.
Unlike MLPs, the same convolution (the same filter values w and b) will be used to find the result for all time stamps \(t\in [1,T]\). This is a very powerful property (called weight sharing) of the CNNs which enables them to learn filters that are invariant across the time dimension.
When considering an MTS as input to a convolutional layer, the filter no longer has one dimension (time) but also has dimensions that are equal to the number of dimensions of the input MTS.
Finally, instead of setting manually the values of the filter \(\omega \), these values should be learned automatically since they depend highly on the targeted dataset. For example, one dataset would have the optimal filter to be equal to [1, 2, 2] whereas another dataset would have an optimal filter equal to \([2,0,1]\). By optimal we mean a filter whose application will enable the classifier to easily discriminate between the dataset classes (see Fig. 2). In order to learn automatically a discriminative filter, the convolution should be followed by a discriminative classifier, which is usually preceded by a pooling operation that can either be local or global.
Local pooling such as average or max pooling takes an input time series and reduces its length T by aggregating over a sliding window of the time series. For example if the sliding window’s length is equal to 3 the resulting pooled time series will have a length equal to \(\frac{T}{3}\)  this is only true if the stride is equal to the sliding window’s length. With a global pooling operation, the time series will be aggregated over the whole time dimension resulting in a single real value. In other words, this is similar to applying a local pooling with a sliding window’s length equal to the length of the input time series. Usually a global aggregation is adopted to reduce drastically the number of parameters in a model thus decreasing the risk of overfitting while enabling the use of CAM to explain the model’s decision (Zhou et al. 2016).
In addition to pooling layers, some deep learning architectures include normalization layers to help the network converge quickly. For time series data, the batch normalization operation is performed over each channel therefore preventing the internal covariate shift across one minibatch training of time series (Ioffe and Szegedy 2015). Another type of normalization was proposed by Ulyanov et al. (2016) to normalize each instance instead of a per batch basis, thus learning the mean and standard deviation of each training instance for each layer via gradient descent. The latter approach is called instance normalization and mimics learning the znormalization parameters for the time series training data.
The final discriminative layer takes the representation of the input time series (the result of the convolutions) and give a probability distribution over the class variables in the dataset. Usually, this layer is comprised of a softmax operation similarly to the MLPs. Note that for some approaches, we would have an additional nonlinear FC layer before the final softmax layer which increases the number of parameters in a network. Finally in order to train and learn the parameters of a deep CNN, the process is identical to training an MLP: a feedforward pass followed by backpropagation (LeCun et al. 1998b). An example of a CNN architecture for TSC with three convolutional layers is illustrated in Fig. 3.
Echo state networks
Another popular type of architectures for deep learning models is the Recurrent Neural Network (RNN). Apart from time series forecasting, we found that these neural networks were rarely applied for time series classification which is mainly due to three factors: (1) the type of this architecture is designed mainly to predict an output for each element (time stamp) in the time series (Längkvist et al. 2014; 2) RNNs typically suffer from the vanishing gradient problem due to training on long time series (Pascanu et al. 2012; 3) RNNs are considered hard to train and parallelize which led the researchers to avoid using them for computational reasons (Pascanu et al. 2013).
Given the aforementioned limitations, a relatively recent type of recurrent architecture was proposed for time series: Echo State Networks (ESNs) (Gallicchio and Micheli 2017). ESNs were first invented by Jaeger and Haas (2004) for time series prediction in wireless communication channels. They were designed to mitigate the challenges of RNNs by eliminating the need to compute the gradient for the hidden layers which reduces the training time of these neural networks thus avoiding the vanishing gradient problem. These hidden layers are initialized randomly and constitutes the reservoir: the core of an ESN which is a sparsely connected random RNN. Each neuron in the reservoir will create its own nonlinear activation of the incoming signal. The interconnected weights inside the reservoir and the input weights are not learned via gradient descent, only the output weights are tuned using a learning algorithm such as logistic regression or Ridge classifier (Hoerl and Kennard 1970).
To better understand the mechanism of these networks, consider an ESN with input dimensionality M, neurons in the reservoir \(N_r\) and an output dimensionality K equal to the number of classes in the dataset. Let \(X(t) \in \mathbb {R}^M\), \(I(t) \in \mathbb {R}^{N_r}\) and \(\hat{Y}(t) \in \mathbb {R}^K\) denote the vectors of the input Mdimensional MTS, the internal (or hidden) state and the output unit activity for time t respectively. Further let \(W_{in} \in \mathbb {R}^{N_r\times M}\) and \(W \in \mathbb {R}^{N_r\times N_r}\) and \(W_{out} \in \mathbb {R}^{C\times N_r}\) denote respectively the weight matrices for the input time series, the internal connections and the output connections as seen in Fig. 4. The internal unit activity I(t) at time t is updated using the internal state at time step \(t1\) and the input time series element at time t. Formally the hidden state can be computed using the following recurrence:
with f denoting an activation function of the neurons, a common choice is \(tanh(\cdot )\) applied elementwise (Tanisaro and Heidemann 2016). The output can be computed according to the following equation:
thus classifying each time series element X(t). Note that ESNs depend highly on the initial values of the reservoir that should satisfy a predetermined hyperparameter: the spectral radius. Fig. 4 shows an example of an ESN with a univariate input time series to be classified into K classes.
Finally, we should note that for all types of DNNs, a set of techniques was proposed by the deep learning community to enhance neural networks’ generalization capabilities. Regularization methods such as l2norm weight decay (Bishop 2006) or Dropout (Srivastava et al. 2014) aim at reducing overfitting by limiting the activation of the neurons. Another popular technique is data augmentation, which tackles the problem of overfitting a small dataset by increasing the number of training instances (Baird 1992). This method consists in cropping, rotating and blurring images which have been shown to improve the DNNs’ performance for computer vision tasks (Zhang et al. 2017). Although two approaches in this survey include a data augmentation technique, the study of its impact on TSC is currently limited (Ismail Fawaz et al. 2018a).
Generative or discriminative approaches
Deep learning approaches for TSC can be separated into two main categories: the generative and the discriminative models [as proposed in Längkvist et al. (2014)]. We further separate these two groups into subgroups which are detailed in the following subsections and illustrated in Fig. 5.
Generative models
Generative models usually exhibit an unsupervised training step that precedes the learning phase of the classifier (Längkvist et al. 2014). This type of network has been referred to as Modelbased classifiers in the TSC community (Bagnall et al. 2017). Some of these generative non deep learning approaches include autoregressive models (Bagnall and Janacek 2014), hidden Markov models (Kotsifakos and Papapetrou 2014) and kernel models (Chen et al. 2013).
For all generative approaches, the goal is to find a good representation of time series prior to training a classifier (Längkvist et al. 2014). Usually, to model the time series, classifiers are preceded by an unsupervised pretraining phase such as stacked denoising autoencoders (SDAEs) (Bengio et al. 2013; Hu et al. 2016). A generative CNNbased model was proposed in Wang et al. (2016c), Mittelman (2015) where the authors introduced a deconvolutional operation followed by an upsampling technique that helps in reconstructing a multivariate time series. Deep Belief Networks (DBNs) were also used to model the latent features in an unsupervised manner which are then leveraged to classify univariate and multivariate time series (Wang et al. 2017a; Banerjee et al. 2017). In Mehdiyev et al. (2017), Malhotra et al. (2018), Rajan and Thiagarajan (2018), an RNN autoencoder was designed to first generate the time series then using the learned latent representation, they trained a classifier (such as SVM or Random Forest) on top of these representations to predict the class of a given input time series.
Other studies such as in Aswolinskiy et al. (2017), Bianchi et al. (2018), Chouikhi et al. (2018), Ma et al. (2016) used selfpredict modeling for time series classification where ESNs were first used to reconstruct the time series and then the learned representation in the reservoir space was utilized for classification. We refer to this type of architecture by traditional ESNs in Fig. 5. Other ESNbased approaches (Chen et al. 2015a, 2013; Che et al. 2017b) define a kernel over the learned representation followed by an SVM or an MLP classifier. In Gong et al. (2018), Wang et al. (2016a), a metalearning evolutionarybased algorithm was proposed to construct an optimal ESN architecture for univariate and multivariate time series. For more details concerning generative ESN models for TSC, we refer the interested reader to a recent empirical study (Aswolinskiy et al. 2016) that compared classification in reservoir and modelspace for both multivariate and univariate time series.
Discriminative models
A discriminative deep learning model is a classifier (or regressor) that directly learns the mapping between the raw input of a time series (or its hand engineered features) and outputs a probability distribution over the class variables in a dataset. Several discriminative deep learning architectures have been proposed to solve the TSC task, but we found that this type of model could be further subdivided into two groups: (1) deep learning models with hand engineered features and (2) endtoend deep learning models.
The most frequently encountered and computer vision inspired feature extraction method for hand engineering approaches is the transformation of time series into images using specific imaging methods such as Gramian fields (Wang and Oates 2015b, a), recurrence plots (Hatami et al. 2017; Tripathy and Acharya 2018) and Markov transition fields (Wang and Oates 2015c). Unlike image transformation, other feature extraction methods are not domain agnostic. These features are first handengineered using some domain knowledge, then fed to a deep learning discriminative classifier. For example in Uemura et al. (2018), several features (such as the velocity) were extracted from sensor data placed on a surgeon’s hand in order to determine the skill level during surgical training. In fact, most of the deep learning approaches for TSC with some hand engineered features are present in human activity recognition tasks (Ignatov 2018). For more details on the different applications of deep learning for human motion detection using mobile and wearable sensor networks, we refer the interested reader to a recent survey (Nweke et al. 2018) where deep learning approaches (with or without hand engineered features) were thoroughly described specifically for the human activity recognition task.
In contrast to feature engineering, endtoend deep learning aims to incorporate the feature learning process while finetuning the discriminative classifier (Nweke et al. 2018). Since this type of deep learning approach is domain agnostic and does not include any domain specific preprocessing steps, we decided to further separate these endtoend approaches using their neural network architectures.
In Wang et al. (2017b), Geng and Luo (2018), an MLP was designed to learn from scratch a discriminative time series classifier. The problem with an MLP approach is that temporal information is lost and the features learned are no longer timeinvariant. This is where CNNs are most useful, by learning spatially invariant filters (or features) from raw input time series (Wang et al. 2017b). During our study, we found that CNN is the most widely applied architecture for the TSC problem, which is probably due to their robustness and the relatively small amount of training time compared to complex architectures such as RNNs or MLPs. Several variants of CNNs have been proposed and validated on a subset of the UCR/UEA archive (Chen et al. 2015b; Bagnall et al. 2017) such as Residual Networks (ResNets) (Wang et al. 2017b; Geng and Luo 2018) which add linear shortcut connections for the convolutional layers potentially enhancing the model’s accuracy (He et al. 2016). In Le Guennec et al. (2016), Cui et al. (2016), Wang et al. (2017b), Zhao et al. (2017), traditional CNNs were also validated on the UCR/UEA archive. More recently in Wang et al. (2018b), the architectures proposed in Wang et al. (2017b) were modified to leverage a filter initialization technique based on the Daubechies 4 Wavelet values (Rowe and Abbott 1995). Outside of the UCR/UEA archive, deep learning has reached stateoftheart performance on several datasets in different domains (Längkvist et al. 2014). For spatiotemporal series forecasting problems, such as meteorology and oceanography, DNNs were proposed in Ziat et al. (2017). Strodthoff and Strodthoff (2019) proposed to detect myocardial infractions from electrocardiography data using deep CNNs. For human activity recognition from wearable sensors, deep learning is replacing the feature engineering approaches (Nweke et al. 2018) where features are no longer handdesigned but rather learned by deep learning models trained through backpropagation. One other type of time series data is present in Electronic Health Records, where a recent generative adversarial network with a CNN (Che et al. 2017a) was trained for risk prediction based on patients historical medical records. In Ismail Fawaz et al. (2018b), CNNs were designed to reach stateoftheart performance for surgical skills identification. Liu et al. (2018) leveraged a CNN model for multivariate and lagfeature characteristics in order to achieve stateoftheart accuracy on the Prognostics and Health Management (PHM) 2015 challenge data. Finally, a recent review of deep learning for physiological signals classification revealed that CNNs were the most popular architecture (Faust et al. 2018) for the considered task. We mention one final type of hybrid architectures that showed promising results for the TSC task on the UCR/UEA archive datasets, where mainly CNNs were combined with other types of architectures such as Gated Recurrent Units (Lin and Runger 2018) and the attention mechanism (Serrà et al. 2018). The reader may have noticed that CNNs appear under Auto Encoders as well as under EndtoEnd learning in Fig. 5. This can be explained by the fact that CNNs when trained as Auto Encoders have a complete different objective function than CNNs that are trained in an endtoend fashion.
Now that we have presented the taxonomy for grouping DNNs for TSC, we introduce in the following section the different approaches that we have included in our experimental evaluation. We also explain the motivations behind the selection of these algorithms.
Approaches
In this section, we start by explaining the reasons behind choosing discriminative endtoend approaches for this empirical evaluation. We then describe in detail the nine different deep learning architectures with their corresponding advantages and drawbacks.
Why discriminative endtoend approaches ?
As previously mentioned in Sect. 2, the main characteristic of a generative model is fitting a time series selfpredictor whose latent representation is later fed into an offtheshelf classifier such as Random Forest or SVM. Although these models do sometimes capture the trend of a time series, we decided to leave these generative approaches out of our experimental evaluation for the following reasons:

This type of method is mainly proposed for tasks other than classification or as part of a larger classification scheme (Bagnall et al. 2017);

The informal consensus in the literature is that generative models are usually less accurate than direct discriminative models (Bagnall et al. 2017; Nguyen et al. 2017);

The implementation of these models is usually more complicated than for discriminative models since it introduces an additional step of fitting a time series generator—this has been considered a barrier with most approaches whose code was not publicly available such as Gong et al. (2018), Che et al. (2017b), Chouikhi et al. (2018), Wang et al. (2017a);

The accuracy of these models depends highly on the chosen offtheshelf classifier which is sometimes not even a neural network classifier (Rajan and Thiagarajan 2018).
Given the aforementioned limitations for generative models, we decided to limit our experimental evaluation to discriminative deep learning models for TSC. In addition of restricting the study to discriminative models, we decided to only consider endtoend approaches, thus further leaving classifiers that incorporate feature engineering out of our empirical evaluation. We made this choice because we believe that the main goal of deep learning approaches is to remove the bias due to manually designed features (Ordón̈ez and Roggen 2016), thus enabling the network to learn the most discriminant useful features for the classification task. This has also been the consensus in the human activity recognition literature, where the accuracy of deep learning methods depends highly on the quality of the extracted features (Nweke et al. 2018). Finally, since our goal is to provide an empirical study of domain agnostic deep learning approaches for any TSC task, we found that it is best to compare models that do not incorporate any domain knowledge into their approach.
As for why we chose the nine approaches (described in the next Section), it is first because among all the discriminative endtoend deep learning models for TSC, we wanted to cover a wide range of architectures such as CNNs, Fully CNNs, MLPs, ResNets, ESNs, etc. Second, since we cannot cover an empirical study of all approaches validated in all TSC domains, we decided to only include approaches that were validated on the whole (or a subset of) the univariate time series UCR/UEA archive (Chen et al. 2015b; Bagnall et al. 2017) and/or on the MTS archive (Baydogan 2015). Finally, we chose to work with approaches that do not try to solve a sub task of the TSC problem such as in Geng and Luo (2018) where CNNs were modified to classify imbalanced time series datasets. To justify this choice, we emphasize that imbalanced TSC problems can be solved using several techniques such as data augmentation (Ismail Fawaz et al. 2018a) and modifying the class weights (Geng and Luo 2018). However, any deep learning algorithm can benefit from this type of modification. Therefore if we did include modifications for solving imbalanced TSC tasks, it would be much harder to determine if it is the choice of the deep learning classifier or the modification itself that improved the accuracy of the model. Another sub task that has been at the center of recent studies is early time series classification (Wang et al. 2016b) where deep CNNs were modified to include an early classification of time series. More recently, a deep reinforcement learning approach was also proposed for the early TSC task (Martinez et al. 2018). For further details, we refer the interested reader to a recent survey on deep learning for early time series classification (Santos and Kern 2017).
Compared approaches
After having presented an overview over the recent deep learning approaches for time series classification, we present the nine architectures that we have chosen to compare in this paper.
Multi layer perceptron
The MLP, which is the most traditional form of DNNs, was proposed in Wang et al. (2017b) as a baseline architecture for TSC. The network contains 4 layers in total where each one is fully connected to the output of its previous layer. The final layer is a softmax classifier, which is fully connected to its previous layer’s output and contains a number of neurons equal to the number of classes in a dataset. All three hidden FC layers are composed of 500 neurons with ReLU as the activation function. Each layer is preceded by a dropout operation (Srivastava et al. 2014) with a rate equal to 0.1, 0.2, 0.2 and 0.3 for respectively the first, second, third and fourth layer. Dropout is one form of regularization that helps in preventing overfitting (Srivastava et al. 2014). The dropout rate indicates the percentage of neurons that are deactivated (set to zero) in a feed forward pass during training.
MLP does not have any layer whose number of parameters is invariant across time series of different lengths (denoted by \(\#invar\) in Table 1) which means that the transferability of the network is not trivial: the number of parameters (weights) of the network depends directly on the length of the input time series.
Fully convolutional neural network
Fully Convolutional Neural Networks (FCNs) were first proposed in Wang et al. (2017b) for classifying univariate time series and validated on 44 datasets from the UCR/UEA archive. FCNs are mainly convolutional networks that do not contain any local pooling layers which means that the length of a time series is kept unchanged throughout the convolutions. In addition, one of the main characteristics of this architecture is the replacement of the traditional final FC layer with a Global Average Pooling (GAP) layer which reduces drastically the number of parameters in a neural network while enabling the use of the CAM (Zhou et al. 2016) that highlights which parts of the input time series contributed the most to a certain classification.
The architecture proposed in Wang et al. (2017b) is first composed of three convolutional blocks where each block contains three operations: a convolution followed by a batch normalization (Ioffe and Szegedy 2015) whose result is fed to a ReLU activation function. The result of the third convolutional block is averaged over the whole time dimension which corresponds to the GAP layer. Finally, a traditional softmax classifier is fully connected to the GAP layer’s output.
All convolutions have a stride equal to 1 with a zero padding to preserve the exact length of the time series after the convolution. The first convolution contains 128 filters with a filter length equal to 8, followed by a second convolution of 256 filters with a filter length equal to 5 which in turn is fed to a third and final convolutional layer composed of 128 filters, each one with a length equal to 3.
We can see that FCN does not hold any pooling nor a regularization operation. In addition, one of the advantages of FCNs is the invariance (denoted by \(\#invar\) in Table 1) in the number of parameters for 4 layers (out of 5) across time series of different lengths. This invariance (due to using GAP) enables the use of a transfer learning approach where one can train a model on a certain source dataset and then finetune it on the target dataset (Ismail Fawaz et al. 2018c).
Residual network
The third and final proposed architecture in Wang et al. (2017b) is a relatively deep Residual Network (ResNet). For TSC, this is the deepest architecture with 11 layers of which the first 9 layers are convolutional followed by a GAP layer that averages the time series across the time dimension. The main characteristic of ResNets is the shortcut residual connection between consecutive convolutional layers. Actually, the difference with the usual convolutions (such as in FCNs) is that a linear shortcut is added to link the output of a residual block to its input thus enabling the flow of the gradient directly through these connections, which makes training a DNN much easier by reducing the vanishing gradient effect (He et al. 2016).
The network is composed of three residual blocks followed by a GAP layer and a final softmax classifier whose number of neurons is equal to the number of classes in a dataset. Each residual block is first composed of three convolutions whose output is added to the residual block’s input and then fed to the next layer. The number of filters for all convolutions is fixed to 64, with the ReLU activation function that is preceded by a batch normalization operation. In each residual block, the filter’s length is set to 8, 5 and 3 respectively for the first, second and third convolution.
Similarly to the FCN model, the layers (except the final one) in the ResNet architecture have an invariant number of parameters across different datasets. That being said, we can easily pretrain a model on a source dataset, then transfer and finetune it on a target dataset without having to modify the hidden layers of the network. As we have previously mentioned and since this type of transfer learning approach can give an advantage for certain types of architecture, we leave the exploration of this area of research for future work. The ResNet architecture proposed by Wang et al. (2017b) is depicted in Fig. 6.
Encoder
Originally proposed by Serrà et al. (2018), Encoder is a hybrid deep CNN whose architecture is inspired by FCN (Wang et al. 2017b) with a main difference where the GAP layer is replaced with an attention layer. In Serrà et al. (2018), two variants of Encoder were proposed: the first approach was to train the model from scratch in an endtoend fashion on a target dataset while the second one was to pretrain this same architecture on a source dataset and then finetune it on a target dataset. The latter approach reached higher accuracy thus benefiting from the transfer learning technique. On the other hand, since almost all approaches can benefit to certain degree from a transfer learning method, we decided to implement only the endtoend approach (training from scratch) which already showed high performance in the author’s original paper.
Similarly to FCN, the first three layers are convolutional with some relatively small modifications. The first convolution is composed of 128 filters of length 5; the second convolution is composed of 256 filters of length 11; the third convolution is composed of 512 filters of length 21. Each convolution is followed by an instance normalization operation (Ulyanov et al. 2016) whose output is fed to the Parametric Rectified Linear Unit (PReLU) (He et al. 2015) activation function. The output of PReLU is followed by a dropout operation (with a rate equal to 0.2) and a final max pooling of length 2. The third convolutional layer is fed to an attention mechanism (Bahdanau et al. 2015) that enables the network to learn which parts of the time series (in the time domain) are important for a certain classification. More precisely, to implement this technique, the input MTS is multiplied with a second MTS of the same length and number of channels, except that the latter has gone through the softmax function. Each element in the second MTS will act as a weight for the first MTS, thus enabling the network to learn the importance of each element (time stamp). Finally, a traditional softmax classifier is fully connected to the latter layer with a number of neurons equal to the number of classes in the dataset.
In addition to replacing the GAP layer with the attention layer, Encoder differs from FCN in three main core changes: (1) the PReLU activation function where an additional parameter is added for each filter to enable learning the slope of the function, (2) the dropout regularization technique and (3) the max pooling operation. One final note is that the careful design of Encoder’s attention mechanism enabled the invariance across all layers which encouraged the authors to implement a transfer learning approach.
Multiscale convolutional neural network
Originally proposed by Cui et al. (2016), Multiscale Convolutional Neural Network (MCNN) is the earliest approach to validate an endtoend deep learning architecture on the UCR Archive. MCNN’s architecture is very similar to a traditional CNN model: with two convolutions (and max pooling) followed by an FC layer and a final softmax layer. On the other hand, this approach is very complex with its heavy data preprocessing step. Cui et al. (2016) were the first to introduce the Window Slicing (WS) method as a data augmentation technique. WS slides a window over the input time series and extract subsequences, thus training the network on the extracted subsequences instead of the raw input time series. Following the extraction of a subsequence from an input time series using the WS method, a transformation stage is used. More precisely, prior to any training, the subsequence will undergo three transformations: (1) identity mapping; (2) downsampling and (3) smoothing; thus, transforming a univariate input time series into a multivariate input time series. This heavy preprocessing would question the endtoend label of this approach, but since their method is generic enough we incorporated it into our developed framework.
For the first transformation, the input subsequence is left unchanged and the raw subsequence will be used as an input for an independent first convolution. The downsampling technique (second transformation) will result in shorter subsequences with different lengths which will then undergo another independent convolutions in parallel to the first convolution. As for the smoothing technique (third transformation), the result is a smoothed subsequence whose length is equal to the input raw subsequence which will also be fed to an independent convolution in parallel to the first and the second convolutions.
The output of each convolution in the first convolutional stage is concatenated to form the input of the subsequent convolutional layer. Following this second layer, an FC layer is deployed with 256 neurons using the sigmoid activation function. Finally, the usual softmax classifier is used with a number of neurons equal to the number of classes in the dataset.
Note that each convolution in this network uses 256 filters with the sigmoid as an activation function, followed by a max pooling operation. Two architecture hyperparameters are crossvalidated, using a grid search on an unseen split from the training set: the filter length and the pooling factor which determines the pooling size for the max pooling operation. The total number of layers in this network is 4, out of which only the first two convolutional layers are invariant (transferable). Finally, since the WS method is also used at test time, the class of an input time series is determined by a majority vote over the extracted subsequences’ predicted labels.
Time LeNet
Time LeNet (tLeNet) was originally proposed by Le Guennec et al. (2016) and inspired by the great performance of LeNet’s architecture for the document recognition task (LeCun et al. 1998a). This model can be considered as a traditional CNN with two convolutions followed by an FC layer and a final softmax classifier. There are two main differences with the FCNs: (1) an FC layer and (2) local maxpooling operations. Unlike GAP, local pooling introduces invariance to small perturbations in the activation map (the result of a convolution) by taking the maximum value in a local pooling window. Therefore for a pool size equal to 2, the pooling operation will halve the length of a time series by taking the maximum value between each two time steps.
For both convolutions, the ReLU activation function is used with a filter length equal to 5. For the first convolution, 5 filters are used and followed by a max pooling of length equal to 2. The second convolution uses 20 filters followed by a max pooling of length equal to 4. Thus, for an input time series of length l, the resulting output of these two convolutions will divide the length of the time series by \(8=4\times 2\). The convolutional blocks are followed by a nonlinear fully connected layer which is composed of 500 neurons, each one using the ReLU activation function. Finally, similarly to all previous architectures, the number of neurons in the final softmax classifier is equal to the number of classes in a dataset.
Unlike ResNet and FCN, this approach does not have much invariant layers (2 out of 4) due to the use of an FC layer instead of a GAP layer, thus increasing drastically the number of parameters needed to be trained which also depends on the length of the input time series. Thus, the transferability of this network is limited to the first two convolutions whose number of parameters depends solely on the number and length of the chosen filters.
We should note that tLeNet is one of the approaches adopting a data augmentation technique to prevent overfitting especially for the relatively small time series datasets in the UCR/UEA archive. Their approach uses two data augmentation techniques: WS and Window Warping (WW). The former method is identical to MCNN’s data augmentation technique originally proposed in Cui et al. (2016). As for the second data augmentation technique, WW employs a warping technique that squeezes or dilates the time series. In order to deal with multilength time series the WS method is adopted to ensure that subsequences of the same length are extracted for training the network. Therefore, a given input time series of length l is first dilated (\(\times \,2\)) then squeezed (\(\times \,\frac{1}{2}\)) resulting in three time series of length l, 2l and \(\frac{1}{2}l\) that are fed to WS to extract equal length subsequences for training. Not that in their original paper (Le Guennec et al. 2016), WS’ length is set to 0.9l. Finally similarly to MCNN, since the WS method is also used at test time, a majority vote over the extracted subsequences’ predicted labels is applied.
Multi channel deep convolutional neural network
Multi Channel Deep Convolutional Neural Network (MCDCNN) was originally proposed and validated on two multivariate time series datasets (Zheng et al. 2014, 2016). The proposed architecture is mainly a traditional deep CNN with one modification for MTS data: the convolutions are applied independently (in parallel) on each dimension (or channel) of the input MTS.
Each dimension for an input MTS will go through two convolutional stages with 8 filters of length 5 with ReLU as the activation function. Each convolution is followed by a max pooling operation of length 2. The output of the second convolutional stage for all dimensions is concatenated over the channels axis and then fed to an FC layer with 732 neurons with ReLU as the activation function. Finally, the softmax classifier is used with a number of neurons equal to the number of classes in the dataset. By using an FC layer before the softmax classifier, the transferability of this network is limited to the first and second convolutional layers.
Time convolutional neural network
TimeCNN approach was originally proposed by Zhao et al. (2017) for both univariate and multivariate TSC. There are three main differences compared to the previously described networks. The first characteristic of TimeCNN is the use of the mean squared error (MSE) instead of the traditional categorical crossentropy loss function, which has been used by all the deep learning approaches we have mentioned so far. Hence, instead of a softmax classifier, the final layer is a traditional FC layer with sigmoid as the activation function, which does not guarantee a sum of probabilities equal to 1. Another difference to traditional CNNs is the use of a local average pooling operation instead of local max pooling. In addition, unlike MCDCNN, for MTS data they apply one convolution for all the dimensions of a multivariate classification task. Another unique characteristic of this architecture is that the final classifier is fully connected directly to the output of the second convolution, which removes completely the GAP layer without replacing it with an FC nonlinear layer.
The network is composed of two consecutive convolutional layers with respectively 6 and 12 filters followed by a local average pooling operation of length 3. The convolutions adopt the sigmoid as the activation function. The network’s output consists of an FC layer with a number of neurons equal to the number of classes in the dataset.
Time warping invariant echo state network
Time Warping Invariant Echo State Network (TWIESN) (Tanisaro and Heidemann 2016) is the only nonconvolutional recurrent architecture tested and reimplemented in our study. Although ESNs were originally proposed for time series forecasting, Tanisaro and Heidemann (2016) proposed a variant of ESNs that uses directly the raw input time series and predicts a probability distribution over the class variables.
In fact, for each element (time stamp) in an input time series, the reservoir space is used to project this element into a higher dimensional space. Thus, for a univariate time series, the element is projected into a space whose dimensions are inferred from the size of the reservoir. Then for each element, a Ridge classifier (Hoerl and Kennard 1970) is trained to predict the class of each time series element. During test time, for each element of an input test time series, the already trained Ridge classifier will output a probability distribution over the classes in a dataset. Then the a posteriori probability for each class is averaged over all time series elements, thus assigning for each input test time series the label for which the averaged probability is maximum. Following the original paper of Tanisaro and Heidemann (2016), using a gridsearch on an unseen split (\(20\%\)) from the training set, we optimized TWIESN’s three hyperparameters: the reservoir’s size, sparsity and spectral radius.
Hyperparameters
Tables 1 and 2 show respectively the architecture and the optimization hyperparameters for all the described approaches except for TWIESN, since its hyperparameters are not compatible with the eight other algorithms’ hyperparameters. We should add that for all the other deep learning classifiers (with TWIESN omitted), a model checkpoint procedure was performed either on the training set or a validation set (split from the training set). Which means that if the model is trained for 1000 epochs, the best one on the validation set (or the train set) loss will be chosen for evaluation. This characteristic is included in Table 2 under the “valid” column. In addition to the model checkpoint procedure, we should note that all deep learning models in Table 1 were initialized randomly using Glorot’s uniform initialization method (Glorot and Bengio 2010). All models were optimized using a variant of Stochastic Gradient Descent (SGD) such as Adam (Kingma and Ba 2015) and AdaDelta (Zeiler 2012). We should add that for FCN, ResNet and MLP proposed in Wang et al. (2017b), the learning rate was reduced by a factor of 0.5 each time the model’s training loss has not improved for 50 consecutive epochs (with a minimum value equal to 0.0001). One final note is that we have no way of controlling the fact that those described architectures might have been overfitted for the UCR/UEA archive and designed empirically to achieve a high performance, which is always a risk when comparing classifiers on a benchmark (Bagnall et al. 2017). We therefore think that challenges where only the training data is publicly available and the testing data are held by the challenge organizer for evaluation might help in mitigating this problem.
Experimental setup
We first start by presenting the datasets’ properties we have adopted in this empirical study. We then describe in details our developed opensource framework of deep learning for time series classification.
Datasets
Univariate archive
In order to have a thorough and fair experimental evaluation of all approaches, we tested each algorithm on the whole UCR/UEA archive (Chen et al. 2015b; Bagnall et al. 2017) which contains 85 univariate time series datasets. The datasets possess different varying characteristics such as the length of the series which has a minimum value of 24 for the ItalyPowerDemand dataset and a maximum equal to 2709 for the HandOutLines dataset. One important characteristic that could impact the DNNs’ accuracy is the size of the training set which varies between 16 and 8926 for respectively DiatomSizeReduction and ElectricDevices datasets. We should note that twenty datasets contains a relatively small training set (50 or fewer instances) which surprisingly was not an impediment for obtaining high accuracy when applying a very deep architecture such as ResNet. Furthermore, the number of classes varies between 2 (for 31 datasets) and 60 (for the ShapesAll dataset). Note that the time series in this archive are already znormalized (Bagnall et al. 2017).
Other than the fact of being publicly available, the choice of validating on the UCR/UEA archive is motivated by having datasets from different domains which have been broken down into seven different categories (Image Outline, Sensor Readings, Motion Capture, Spectrographs, ECG, Electric Devices and Simulated Data) in Bagnall et al. (2017). Further statistics, which we do not repeat for brevity, were conducted on the UCR/UEA archive in Bagnall et al. (2017).
Multivariate archive
We also evaluated all deep learning models on Baydogan’s archive (Baydogan 2015) that contains 13 MTS classification datasets. For memory usage limitations over a single GPU, we left the MTS dataset Performance Measurement System (PeMS) out of our experimentations. This archive also exhibits datasets with different characteristics such as the length of the time series which, unlike the UCR/UEA archive, varies among the same dataset. This is due to the fact that the datasets in the UCR/UEA archive are already rescaled to have an equal length among one dataset (Bagnall et al. 2017).
In order to solve the problem of unequal length time series in the MTS archive we decided to linearly interpolate the time series of each dimension for every given MTS, thus each time series will have a length equal to the longest time series’ length. This form of preprocessing has also been used by Ratanamahatana and Keogh (2005) to show that the length of a time series is not an issue for TSC problems. This step is very important for deep learning models whose architecture depends on the length of the input time series (such as a MLP) and for parallel computation over the GPUs. We did not znormalize any time series, but we emphasize that this traditional preprocessing step (Bagnall et al. 2017) should be further studied for univariate as well as multivariate data, especially since normalization is known to have a huge effect on DNNs’ learning capabilities (Zhang et al. 2017). Note that this process is only true for the MTS datasets whereas for the univariate benchmark, the time series are already znormalized. Since the data is preprocessed using the same technique for all nine classifiers, we can safely say, to some extent, that the accuracy improvement of certain models can be solely attributed to the model itself. Table 3 shows the different characteristics of each MTS dataset used in our experiments.
Experiments
For each dataset in both archives (97 datasets in total), we have trained the nine deep learning models (presented in the previous Section) with 10 different runs each. Each run uses the same original train/test split in the archive but with a different random weight initialization, which enables us to take the mean accuracy over the 10 runs in order to reduce the bias due to the weights’ initial values. In total, we have performed 8730 experiments for the 85 univariate and 12 multivariate TSC datasets. Thus, given the huge number of models that needed to be trained, we ran our experiments on a cluster of 60 GPUs. These GPUs were a mix of four types of Nvidia graphic cards: GTX 1080 Ti, Tesla K20, K40 and K80. The total sequential running time was approximately 100 days, that is if the computation has been done on a single GPU. However, by leveraging the cluster of 60 GPUs, we managed to obtain the results in less than one month. We implemented our framework using the open source deep learning library Keras (Chollet 2015) with the Tensorflow (Abadi et al. 2015) backend^{Footnote 1}.
Following Lucas et al. (2018), Forestier et al. (2017), Petitjean et al. (2016), Grabocka et al. (2014) we used the mean accuracy measure averaged over the 10 runs on the test set. When comparing with the stateoftheart results published in Bagnall et al. (2017) we averaged the accuracy using the median test error. Following the recommendation in Demšar (2006) we used the Friedman test (Friedman 1940) to reject the null hypothesis. Then we performed the pairwise posthoc analysis recommended by Benavoli et al. (2016) where the average rank comparison is replaced by a Wilcoxon signedrank test (Wilcoxon 1945) with Holm’s alpha (\(5\%\)) correction (Holm 1979; Garcia and Herrera 2008). To visualize this type of comparison we used a critical difference diagram proposed by Demšar (2006), where a thick horizontal line shows a group of classifiers (a clique) that are notsignificantly different in terms of accuracy.
Results
In this section, we present the accuracies for each one of the nine approaches. All accuracies are absolute and not relative to each other that is if we claim algorithm A is 5% better than algorithm B, this means that the average accuracy is 0.05 higher for algorithm A than B.
Results for univariate time series
We provide on the companion GitHub repository the raw accuracies over the 10 runs for the nine deep learning models we have tested on the 85 univariate time series datasets: the UCR/UEA archive (Chen et al. 2015b; Bagnall et al. 2017). The corresponding critical difference diagram is shown in Fig. 7. The ResNet significantly outperforms the other approaches with an average rank of almost 2. ResNet wins on 50 problems out of 85 and significantly outperforms the FCN architecture. This is in contrast to the original paper’s results where FCN was found to outperform ResNet on 18 out of 44 datasets, which shows the importance of validating on a larger archive in order to have a robust statistical significance.
We believe that the success of ResNet is highly due to its deep flexible architecture. First of all, our findings are in agreement with the deep learning for computer vision literature where deeper neural networks are much more successful than shallower architectures (He et al. 2016). In fact, in a space of 4 years, neural networks went from 7 layers in AlexNet 2012 (Krizhevsky et al. 2012) to 1000 layers for ResNet 2016 (He et al. 2016). These types of deep architectures generally need a huge amount of data in order to generalize well on unseen examples (He et al. 2016). Although the datasets used in our experiments are relatively small compared to the billions of labeled images (such as ImageNet (Russakovsky et al. 2015) and OpenImages (Krasin et al. 2017) challenges), the deepest networks did reach competitive accuracies on the UCR/UEA archive benchmark.
We give two potential reasons for this high generalization capabilities of deep CNNs on the TSC tasks. First, having seen the success of convolutions in classification tasks that require learning features that are spatially invariant in a two dimensional space (such as width and height in images), it is only natural to think that discovering patterns in a one dimensional space (time) should be an easier task for CNNs thus requiring less data to learn from. The other more direct reason behind the high accuracies of deep CNNs on time series data is its success in other sequential data such as speech recognition (Hinton et al. 2012) and sentence classification (Kim 2014) where text and audio, similarly to time series data, exhibit a natural temporal ordering.
The MCNN and tLeNet architectures yielded very low accuracies with only one win for the Earthquakes dataset. The main common idea between both of these approaches is extracting subsequences to augment the training data. Therefore the model learns to classify a time series from a shorter subsequence instead of the whole one, then with a majority voting scheme the time series at test time are assigned a class label. The poor performances (worst average ranks) for these two approaches suggest that this adhoc method of slicing the time series does not guarantee that the discriminative information of a time series has not been lost. These two classifiers are similar to the phase dependent intervals TSC algorithms (Bagnall et al. 2017) where the classifiers derive features from intervals of each series. Similarly to the recent comparative study of TSC algorithms, this type of Window Slicing based approach yielded the lowest average ranks.
Although MCDCNN and TimeCNN were originally proposed to classify MTS datasets, we have evaluated them on the univariate UCR/UEA archive. The MCDCNN did not manage to beat any of the classifiers except for the ECG5000 dataset which is already a dataset where almost all approaches reached the highest accuracy. This low performance is probably due to the nonlinear FC layer that replaces the GAP pooling of the best performing algorithms (FCN and ResNet). This FC layer reduces the effect of learning time invariant features which explains why MLP, TimeCNN and MCDCNN exhibit very similar performance.
One approach that shows relatively high accuracy is Encoder (Serrà et al. 2018). The statistical test indicates a significant difference between Encoder, FCN and ResNet. FCN wins on 36 datasets whereases Encoder wins only on 17 which suggests the superiority of the GAP layer compared to Encoder’s attention mechanism.
Comparing with stateoftheart approaches
In this section, we compared ResNet (the most accurate DNN of our study) with the current stateoftheart classifiers evaluated on the UCR/UEA archive in the great time series classification bake off (Bagnall et al. 2017). Note that our empirical study strongly suggests to use ResNet instead of any other deep learning algorithm—it is the most accurate one with similar runtime to FCN (the second most accurate DNN). Finally, since ResNet’s results were averaged over ten different random initializations, we chose to take one iteration of ResNet (the median) and compare it to other stateoftheart algorithms that were executed once over the original train/test split provided by the UCR/UEA archive.
Out of the 18 classifiers evaluated by Bagnall et al. (2017), we have chosen the four best performing algorithms: (1) Elastic Ensemble (EE) proposed by Lines and Bagnall (2015) is an ensemble of nearest neighbor classifiers with 11 different time series similarity measures; (2) BagofSFASymbols (BOSS) published in Schäfer (2015) forms a discriminative bag of words by discretizing the time series using a Discrete Fourier Transform and then building a nearest neighbor classifier with a bespoke distance measure; (3) Shapelet Transform (ST) developed by Hills et al. (2014) extracts discriminative subsequences (shapelets) and builds a new representation of the time series that is fed to an ensemble of 8 classifiers; (4) Collective of Transformationbased Ensembles (COTE) proposed by Bagnall et al. (2017) is basically a weighted ensemble of 35 TSC algorithms including EE and ST. We also include the Hierarchical Vote Collective of TransformationBased Ensembles (HIVECOTE) proposed by Lines et al. (2018) which improves significantly COTE’s performance by leveraging a hierarchical voting system as well as adding two new classifiers and two additional transformation domains. In addition to these five stateoftheart classifiers, we have included the classic nearest neighbor coupled with DTW and a warping window (WW) set through crossvalidation on the training set (denoted by NNDTWWW), since it is still one of the most popular methods for classifying time series data (Bagnall et al. 2017). Finally, we added a recent approach named Proximity Forest (PF) which is similar to Random Forest but replaces the attribute based splitting criteria by a random similarity measure chosen out of EE’s elastic distances (Lucas et al. 2018). Note that we did not implement any of the nondeep TSC algorithms. We used the results provided by Bagnall et al. (2017) and the other corresponding papers to construct the critical difference diagram in Fig. 8.
Figure 8 shows the critical difference diagram over the UEA benchmark with ResNet added to the pool of six classifiers. As we have previously mentioned, the stateoftheart classifiers are compared to ResNet’s median accuracy over the test set. Nevertheless, we generated the ten different average ranks for each iteration of ResNet and observed that the ranking of the compared classifiers is stable for the ten different random initializations of ResNet. The statistical test failed to find any significant difference between COTE/HIVECOTE and ResNet which is the only TSC algorithm that was able to reach similar performance to COTE. Note that for the ten different random initializations of ResNet, the pairwise statistical test always failed to find any significance between ResNet and COTE/HIVECOTE. PF, ST, BOSS and ResNet showed similar performances according to the Wilcoxon signedrank test, but the fact that ResNet is not significantly different than COTE suggests that more datasets would give a better insight into these performances (Demšar 2006). NNDTWWW and EE showed the lowest average rank suggesting that these methods are no longer competitive with current stateoftheart algorithms for TSC. It is worthwhile noting that cliques formed by the Wilcoxon Signed Rank Test with Holm’s alpha correction do not necessary reflect the rank order (Lines et al. 2018). For example, if we have three classifiers (\(C_1,C_2,C_3\)) with average ranks (\(C_1>C_2>C_3\)), one can still encounter a case where \(C_1\) is not significantly worse than \(C_2\) and \(C_3\) with \(C_2\) and \(C_3\) being significantly different. In our experiments, when comparing to stateoftheart algorithms, we have encountered this problem with (ResNet > COTE > HIVECOTE). Therefore we should emphasize that HIVECOTE and COTE are significantly different when performing the pairwise statistical test.
Although HIVECOTE is still the most accurate classifier (when evaluated on the UCR/UEA archive) its use in a real data mining application is limited due to its huge training time complexity which is \(O(N^2\cdot T^4)\) corresponding to the training time of one of its individual classifiers ST. However, we should note that the recent work of Bostrom and Bagnall (2015) showed that it is possible to use a random sampling approach to decrease significantly the running time of ST (HIVECOTE’s chokepoint) without any loss of accuracy. On the other hand, DNNs offer this type of scalability evidenced by its revolution in the field of computer vision when applied to images, which are thousand times larger than time series data (Russakovsky et al. 2015). In addition to the huge training time, HIVECOTE’s classification time is bounded by a linear scan of the training set due to employing a nearest neighbor classifier, whereas the trivial GPU parallelization of DNNs provides instant classification. Finally we should note that unlike HIVECOTE, ResNet’s hyperparameters were not tuned for each dataset but rather the same architecture was used for the whole benchmark suggesting further investigation of these hyperparameters should improve DNNs’ accuracy for TSC. These results should give an insight of deep learning for TSC therefore encouraging researchers to consider the DNNs as robust real time classifiers for time series data.
The need of a fair comparison
In this section, we highlight the fairness of the comparison to other machine learning TSC algorithms. Since we did not train nor test any of the stateoftheart non deep learning algorithms, it is possible that we allowed much more training time for the described DNNs. For example, for a lazy machine learning algorithm such as NNDTW, training time is zero when allowing maximum warping whereas it has been shown that judicially setting the warping window (Dau et al. 2017) can lead to a significant increase in accuracy. Therefore, we believe that allowing a much more thorough search of DTW’s warping window would lead to a fairer comparison between deep learning approaches and other stateoftheart TSC algorithms. In addition to crossvalidating NNDTW’s hyperparameters, we can imagine spending more time on data preprocessing and cleansing (e.g. smoothing the input time series) in order to improve the accuracy of NNDTW (Höppner 2016; Dau et al. 2018). Ultimately, in order to obtain a fair comparison between deep learning and current stateoftheart algorithms for TSC, we think that the time spent on optimizing a network’s weights should be also spent on optimizing non deep learning based classifiers especially lazy learning algorithms such as the K nearest neighbor coupled with any similarity measure.
Results for multivariate time series
We provide on our companion repository^{Footnote 2} the detailed performance of the nine deep learning classifiers for 10 different random initializations over the 12 MTS classification datasets (Baydogan 2015). Although TimeCNN and MCDCNN are the only architectures originally proposed for MTS data, they were outperformed by the three deep CNNs (ResNet, FCN and Encoder), which shows the superiority of these approaches on the MTS classification task. The corresponding critical difference diagram is depicted in Fig. 9, where the statistical test failed to find any significant difference between the nine classifiers which is mainly due to the small number of datasets compared to their univariate counterpart. Therefore, we illustrated in Fig. 10 the critical difference diagram when both archives are combined (evaluation on 97 datasets in total). At first glance, we can notice that when adding the MTS datasets to the evaluation, the critical difference diagram in Fig. 10 is not significantly different than the one in Fig. 7 (where only the univariate UCR/UEA archive was taken into consideration). This is probably due to the fact that the algorithms’ performance over the 12 MTS datasets is negligible to a certain degree when compared to the performance over the 85 univariate datasets. These observations reinforces the need to have an equally large MTS classification archive in order to evaluate hybrid univariate/multivariate time series classifiers. The rest of the analysis is dedicated to studying the effect of the datasets’ characteristics on the algorithms’ performance.
What can the dataset’s characteristics tell us about the best architecture?
The first dataset characteristic we have investigated is the problem’s domain. Table 4 shows the algorithms’ performance with respect to the dataset’s theme. These themes were first defined in Bagnall et al. (2017). Again, we can clearly see the dominance of ResNet as the best performing approach across different domains. One exception is the electrocardiography (ECG) datasets (7 in total) where ResNet was drastically beaten by the FCN model in 71.4% of ECG datasets. However, given the small sample size (only 7 datasets), we cannot conclude that FCN will almost always outperform the ResNet model for ECG datasets (Bagnall et al. 2017).
The second characteristic which we have studied is the time series length. Similar to the findings for non deep learning models in Bagnall et al. (2017), the time series length does not give information on deep learning approaches’ performance. Table 5 shows the average rank of each DNN over the univariate datasets grouped by the datasets’ lengths. One might expect that the relatively short filters (3) might affect the performance of ResNet and FCN since longer patterns cannot be captured by short filters. However, since increasing the number of convolutional layers will increase the path length viewed by the CNN model (Vaswani et al. 2017), ResNet and FCN managed to outperform other approaches whose filter length is longer (21) such as Encoder. For the recurrent TWIESN algorithm, we were expecting a poor accuracy for very long time series since a recurrent model may “forget” a useful information present in the early elements of a long time series. However, TWIESN did reach competitive accuracies on several long time series datasets such as reaching a 96.8% accuracy on Meat whose time series length is equal to 448. This would suggest that ESNs can solve the vanishing gradient problem especially when learning from long time series.
A third important characteristic is the training size of datasets and how it affects a DNN’s performance. Table 6 shows the average rank for each classifier grouped by the train set’s size. Again, ResNet and FCN still dominate with not much of a difference. However we found one very interesting dataset: DiatomSizeReduction. ResNet and FCN achieved the worst accuracy (30%) on this dataset while TimeCNN reached the best accuracy (95%). Interestingly, DiatomSizeReduction is the smallest datasets in the UCR/UEA archive (with 16 training instances), which suggests that ResNet and FCN are easily overfitting this dataset. This suggestion is also supported by the fact that TimeCNN is the smallest model: it contains a very small number of parameters by design with only 18 filters compared to the 512 filters of FCN. This simple architecture of TimeCNN renders overfitting the dataset much harder. Therefore, we conclude that the small number of filters in TimeCNN is the main reason behind its success on small datasets, however this shallow architecture is unable to capture the variability in larger time series datasets which is modeled efficiently by the FCN and ResNet architectures. One final observation that is in agreement with the deep learning literature is that in order to achieve high accuracies while training a DNN, a large training set is needed. Figure 11 shows the effect of the training size on ResNet’s accuracy for the TwoPatterns dataset: the accuracy increases significantly when adding more training instances until it reaches 100% for 75% of the training data.
Finally, we should note that the number of classes in a dataset—although it yielded some variability in the results for the recent TSC experimental study conducted by Bagnall et al. (2017)—did not show any significance when comparing the classifiers based on this characteristic. In fact, most DNNs architectures, with the categorical crossentropy as their cost function, employ mainly the same classifier: softmax which is basically designed for multiclass classification.
Overall, our results show that, on average, ResNet is the best architecture with FCN and Encoder following as second and third respectively. ResNet performed very well in general except for the ECG datasets where it was outperformed by FCN. MCNN and tLeNet, where time series were cropped into subsequences, were the worst on average. We found small variance between the approaches that replace the GAP layer with an FC dense layer (MCDCNN, CNN) which also showed similar performance to TWIESN and MLP.
Effect of random initializations
The initialization of deep neural networks has received a significant amount of interest from many researchers in the field (LeCun et al. 2015). These advancement have contributed to a better understanding and initialization of deep learning models in order to maximize the quality of nonoptimal solutions found by the gradient descent algorithm (Glorot and Bengio 2010). Nevertheless, we observed in our experiments, that DNNs for TSC suffer from a significant decrease (increase) in accuracy when initialized with bad (good) random weights. Therefore, we study in this section, how random initializations can affect the performance of ResNet and FCN on the whole benchmark in a best and worst case scenario.
Figure 12 shows the accuracy plot of ResNet versus FCN on the 85 univariate time series datasets when aggregated over the 10 random initializations using three different functions: the minimum, median and maximum. When first observing Fig. 12 one can easily conclude that ResNet has a better performance than FCN across most of the datasets regardless of the aggregation method. This is in agreement with the critical difference diagram as well as the analysis conducted in the previous subsections, where ResNet was shown to achieve higher performance on most datasets with different characteristics. A deeper look into the minimum aggregation (red points in Fig. 12) shows that FCN’s performance is less stable compared to ResNet’s. In other words, the weight’s initial value can easily decrease the accuracy of FCN whereas ResNet maintained a relatively high accuracy when taking the worst initial weight values. This is also in agreement with the average standard deviation of ResNet (1.48) which is less than FCN’s (1.70). These observations would encourage a practitioner to avoid using a complex deep learning model since its accuracy may be unstable. Nevertheless we think that investigating different weight initialization techniques such as leveraging the weights of a pretrained neural network would yield better and much more stable results (Ismail Fawaz et al. 2018c).
Visualization
In this section, we start by investigating the use of Class Activation Map to provide an interpretable feedback that highlights the reason for a certain decision taken by the classifier. We then propose another visualization technique which is based on MultiDimensional Scaling (Kruskal and Wish 1978) to understand the latent representation that is learned by the DNNs.
Class activation map
We investigate the use of Class Activation Map (CAM) which was first introduced by Zhou et al. (2016) to highlight the parts of an image that contributed the most for a given class identification. Wang et al. (2017b) later introduced a onedimensional CAM with an application to TSC. This method explains the classification of a certain deep learning model by highlighting the subsequences that contributed the most to a certain classification. Figures 13 and 14 show the results of applying CAM respectively on GunPoint and Meat datasets. Note that employing the CAM is only possible for the approaches with a GAP layer preceding the softmax classifier (Zhou et al. 2016). Therefore, we only considered in this section the ResNet and FCN models, who also achieved the best accuracies overall. Note that Wang et al. (2017b) was the only paper to propose an interpretable analysis of TSC with a DNN. We should emphasize that this is a very important research area which is usually neglected for the sake of improving accuracy: only 2 out of the 9 approaches provided a method that explains the decision taken by a deep learning model. In this section, we start by presenting the CAM method from a mathematical point of view and follow it with two interesting case studies on Meat and GunPoint datasets.
By employing a Global Average Pooling (GAP) layer, ResNet and FCN benefit from the CAM method (Zhou et al. 2016), which makes it possible to identify which regions of an input time series constitute the reason for a certain classification. Formally, let A(t) be the result of the last convolutional layer which is an MTS with M variables. \(A_m(t)\) is the univariate time series for the variable \(m\in [1,M]\), which is in fact the result of applying the \(m\mathrm{th}\) filter. Now let \(w_m^{c}\) be the weight between the \(m\mathrm{th}\) filter and the output neuron of class c. Since a GAP layer is used then the input to the neuron of class c (\(z_c\)) can be computed by the following equation:
The second sum constitutes the averaged time series over the whole time dimension but with the denominator omitted for simplicity. The input \(z_c\) can be also written by the following equation:
Finally the Class Activation Map (\(CAM_c\)) that explains the classification as label c is given in the following equation:
CAM is actually a univariate time series where each element (at time stamp \(t\in [1,T]\)) is equal to the weighted sum of the M data points at t, with the weights being learned by the neural network.
GunPoint dataset
The GunPoint dataset was first introduced by Ratanamahatana and Keogh (2005) as a TSC problem. This dataset involves one male and one female actor performing two actions (GunDraw and Point) which makes it a binary classification problem. For GunDraw (Class1 in Fig. 13), the actors have first their hands by their sides, then draw a replicate gun from hipmounted holster, point it towards the target for one second, then finally place the gun in the holster and their hands to their initial position. Similarly to GunDraw, for Point (Class2 in Fig. 13) the actors follow the same steps but instead of pointing a gun they point their index finger. For each task, the centroid of the actor’s right hands on both X and Y axes were tracked and seemed to be very correlated, therefore the dataset contains only one univariate time series: the Xaxis.
We chose to start by visualizing the CAM for GunPoint for three main reasons. First, it is easy to visualize unlike other noisy datasets. Second, both FCN and ResNet models achieved almost \(100\%\) accuracy on this dataset which will help us to verify if both models are reaching the same decision for the same reasons. Finally, it contains only two classes which allow us to analyze the data much more easily.
Figure 13 shows the CAM’s result when applied on each time series from both classes in the training set while classifying using the FCN model (Fig. 13a, b) and the ResNet model (Fig. 13c, d). At first glance, we can clearly see how both DNNs are neglecting the plateau nondiscriminative regions of the time series when taking the classification decision. It is depicted by the blue flat parts of the time series which indicates no contribution to the classifier’s decision. As for the highly discriminative regions (the red and yellow regions) both models were able to select the same parts of the time series which correspond to the points with high derivatives. Actually, the first most distinctive part of class1 discovered by both classifiers is almost the same: the little red bump in the bottom left of Fig. 13a and c. Finally, another interesting observation is the ability of CNNs to localize a given discriminative shape regardless where it appears in the time series, which is evidence for CNNs’ capability of learning timeinvariant warped features.
An interesting observation would be to compare the discriminative regions identified by a deep learning model with the most discriminative shapelets extracted by other shapeletbased approaches. This observation would also be backed up by the mathematical proof provided by Cui et al. (2016), that showed how the learned filters in a CNN can be considered a generic form of shapelets extracted by the learning shapelets algorithm (Grabocka et al. 2014). Ye and Keogh (2011) identified that the most important shapelet for the Gun/NoGun classification occurs when the actor’s arm is lowered (about 120 on the horizontal axis in Fig. 13). Hills et al. (2014) introduced a shapelet transformation based approach that discovered shapelets that are similar to the ones identified by Ye and Keogh (2011). For ResNet and FCN, the part where the actor lowers his/her arm (bottom right of Fig. 13) seems to be also identified as potential discriminative regions for some time series. On the other hand, the part where the actor raises his/her arm seems to be also a discriminative part of the data which suggests that the deep learning algorithms are identifying more “shapelets”. We should note that this observation cannot confirm which classifier extracted the most discriminative subsequences especially because all algorithms achieved similar accuracy on GunPoint dataset. Perhaps a bigger dataset might provide a deeper insight into the interpretability of these machine learning models. Finally, we stress that the shapelet transformation classifier (Hills et al. 2014) is an ensemble approach, which makes unclear how the shapelets affect the decision taken by the individual classifiers whereas for an endtoend deep learning model we can directly explain the classification by using the Class Activation Map.
Meat dataset
Although the previous case study on GunPoint yielded interesting results in terms of showing that both models are localizing meaningful features, it failed to show the difference between the two most accurate deep learning classifiers: ResNet and FCN. Therefore we decided to further analyze the CAM’s result for the two models on the Meat dataset.
Meat is a food spectrograph dataset which are usually used in chemometrics to classify food types, a task that has obvious applications in food safety and quality assurance. There are three classes in this dataset: Chicken, Pork and Turkey corresponding respectively to classes 1, 2 and 3 in Fig. 14. AlJowder et al. (1997) described how the data is acquired from 60 independent samples using Fourier transform infrared (FTIR) spectroscopy with attenuated total reflectance (ATR) sampling.
Similarly to GunPoint, this dataset is easy to visualize and does not contain very noisy time series. In addition, with only three classes, the visualization is possible to understand and analyze. Finally, unlike for the GunPoint dataset, the two approaches ResNet and FCN reached significantly different results on Meat with respectively \(97\%\) and \(83\%\) accuracy.
Figure 14 enables the comparison between FCN’s CAM (left) and ResNet’s CAM (right). We first observe that ResNet is much more firm when it comes to highlighting the regions. In other words, FCN’s CAM contains much more smoother regions with cyan, green and yellow regions, whereas ResNet’s CAM contains more dark red and blue subsequences showing that ResNet can filter out nondiscriminative and discriminative regions with a higher confidence than FCN, which probably explains why FCN is less accurate than ResNet on this dataset. Another interesting observation is related to the red subsequence highlighted by FCN’s CAM for class 2 and 3 at the bottom right of Fig. 14c and e. By visually investigating this part of the time series, we clearly see that it is a nondiscriminative part since the time series of both classes exhibit this bump. This subsequence is therefore filteredout by the ResNet model which can be seen by the blue color in the bottom right of Fig. 14d and f. These results suggest that ResNet’s superiority over FCN is mainly due to the former’s ability to filterout nondistinctive regions of the time series. We attribute this ability to the main characteristic of ResNet which is composed of the residual connections between the convolutional blocks that enable the model to learn to skip unnecessary convolutions by dint of its shortcut links (He et al. 2016).
Multidimensional scaling
We propose the use of MultiDimensional Scaling (MDS) (Kruskal and Wish 1978) with the objective to gain some insights on the spatial distribution of the input time series belonging to different classes in the dataset. MDS uses a pairwise distance matrix as input and aims at placing each object in a Ndimensional space such as the betweenobject distances are preserved as well as possible. Using the Euclidean Distance (ED) on a set of input time series belonging to the test set, it is then possible to create a similarity matrix and apply MDS to display the set into a two dimensional space. This straightforward approach supposes that the ED is able to strongly separate the raw time series, which is usually not the case evident by the low accuracy of the nearest neighbor when coupled with ED (Bagnall et al. 2017).
On the other hand, we propose to apply this MDS method to visualize the set of time series with its latent representation learned by the network. Usually in a deep neural network, we have several hidden layers and one can find several latent representation of the dataset. But since we are aiming at visualizing the class specific latent space, we chose to use the last latent representation of a DNN (the one directly before the softmax classifier), which is known to be a class specific layer (Yosinski et al. 2014). We decided to apply this method only on ResNet and FCN for two reasons: (1) when evaluated on the UCR/UEA archive they reached the highest ranks; (2) they both employ a GAP layer before the softmax layer making the number of latent features invariant to the time series length.
To better explain this process, for each input time series, the last convolution (for ResNet and FCN) outputs a multivariate time series whose dimensions are equal to the number of filters (128) in the last convolution, then the GAP layer averages the latter 128dimensional multivariate time series over the time dimension resulting in a vector of 128 real values over which the ED is computed. As we worked with the ED, we used metric MDS (Kruskal and Wish 1978) that minimizes a cost function called Stress which is a residual sum of squares:
where \(d_{ij}\) is the ED between the GAP vectors of time series \(X_i\) and \(X_j\). Obviously, one has to be careful about the interpretation of MDS output, as the data space is highly simplified (each time series \(X_i\) is represented as a single data point \(x_i\)).
Figure 15 shows three MDS plots for the GunPoint dataset using: (1) the raw input time series (Fig. 15a); (2) the learned latent features from the GAP layer for FCN (Fig. 15b); and (3) the learned latent features from the GAP layer for ResNet (Fig. 15c). We can easily observe in Fig. 15a that when using the raw input data and projecting it into a 2D space, the two classes are not linearly separable. On the other hand, in both Figs. 15b and c, by applying MDS on the latent representation learned by the network, one can easily separate the set of time series belonging to the two classes. We note that both deep learning models (FCN and ResNet) managed to project the data from GunPoint into a linearly separable space which explains why both models performed equally very well on this dataset with almost 100% accuracy.
Although the visualization of MDS on GunPoint yielded some interesting results, it failed to pinpoint the difference between the two deep learning models FCN and ResNet. Therefore we decided to analyze another dataset where the accuracy of both models differed by almost 15%. Figure 16 shows three MDS plots for the Wine dataset using: (1) the raw input time series (Fig. 16a); (2) the learned latent features from the GAP layer for FCN (Fig. 16b); and (3) the learned latent features from the GAP layer for ResNet (Fig. 16c). At first glimpse of Fig. 16, the reader can conclude that all projections, even when using the learned representation, are not linearly separable which is evident by the relatively low accuracy of both models FCN and ResNet which is equal respectively to 58.7% and 74.4%. A thorough observation shows us that the learned hidden representation of ResNet (Fig. 16c) separates the data from both classes in a much clearer way than the FCN (Fig. 16b). In other words, FCN’s learned representation has too many data points close to the decision boundary whereas ResNet’s hidden features enables projecting data points further away from the decision boundary. This observation could explain why ResNet achieves a better performance than FCN on the Wine dataset.
Conclusion
In this paper, we presented the largest empirical study of DNNs for TSC. We described the most recent successful deep learning approaches for TSC in many different domains such as human activity recognition and sleep stage identification. Under a unified taxonomy, we explained how DNNs are separated into two main categories of generative and discriminative models. We reimplemented nine recently published endtoend deep learning classifiers in a unique framework which we make publicly available to the community. Our results show that endtoend deep learning can achieve the current stateoftheart performance for TSC with architectures such as Fully Convolutional Neural Networks and deep Residual Networks. Finally, we showed how the blackbox effect of deep models which renders them uninterpretable, can be mitigated with a Class Activation Map visualization that highlights which parts of the input time series, contributed the most to a certain class identification.
Although we have conducted an extensive experimental evaluation, deep learning for time series classification, unlike for computer vision and NLP tasks, still lacks a thorough study of data augmentation (Ismail Fawaz et al. 2018a; Forestier et al. 2017) and transfer learning (Ismail Fawaz et al. 2018c; Serrà et al. 2018). In addition, the time series community would benefit from an extension of this empirical study that compares in addition to accuracy, the training and testing time of these deep learning models. Furthermore, we think that the effect of znormalization (and other normalization methods) on the learning capabilities of DNNs should also be thoroughly explored. In our future work, we aim to investigate and answer the aforementioned limitations by conducting more extensive experiments especially on multivariate time series datasets. In order to achieve all of these goals, one important challenge for the TSC community is to provide one large generic labeled dataset similar to the large images database in computer vision such as ImageNet (Russakovsky et al. 2015) that contains 1000 classes.
In conclusion, with data mining repositories becoming more frequent, leveraging deeper architectures that can learn automatically from annotated data in an endtoend fashion, makes deep learning a very enticing approach.
Notes
 1.
The implementations are available on https://github.com/hfawaz/dl4tsc.
 2.
References
Abadi M, Agarwal A, Barham P, Brevdo E, Chen Z, Citro C, Corrado GS, Davis A, Dean J, Devin M, Ghemawat S, Goodfellow I, Harp A, Irving G, Isard M, Jia Y, Jozefowicz R, Kaiser L, Kudlur M, Levenberg J, Mané D, Monga R, Moore S, Murray D, Olah C, Schuster M, Shlens J, Steiner B, Sutskever I, Talwar K, Tucker P, Vanhoucke V, Vasudevan V, Viégas F, Vinyals O, Warden P, Wattenberg M, Wicke M, Yu Y, Zheng X (2015) TensorFlow: largescale machine learning on heterogeneous systems. https://www.tensorflow.org/. Accessed 28 Feb 2019
AlJowder O, Kemsley E, Wilson R (1997) Midinfrared spectroscopy and authenticity problems in selected meats: a feasibility study. Food Chem 59(2):195–201
Aswolinskiy W, Reinhart RF, Steil J (2016) Time series classification in reservoir and modelspace: a comparison. In: Artificial neural networks in pattern recognition, pp 197–208
Aswolinskiy W, Reinhart RF, Steil J (2017) Time series classification in reservoir and modelspace. Neural Process Lett 48:789–809
Bagnall A, Janacek G (2014) A run length transformation for discriminating between auto regressive time series. J Classif 31(2):154–178
Bagnall A, Lines J, Hills J, Bostrom A (2016) Timeseries classification with COTE: the collective of transformationbased ensembles. In: International conference on data engineering, pp 1548–1549
Bagnall A, Lines J, Bostrom A, Large J, Keogh E (2017) The great time series classification bake off: a review and experimental evaluation of recent algorithmic advances. Data Min Knowl Discov 31(3):606–660
Bahdanau D, Cho K, Bengio Y (2015) Neural machine translation by jointly learning to align and translate. In: International conference on learning representations
Baird HS (1992) Document image defect models. Springer, Berlin, pp 546–556
Banerjee D, Islam K, Mei G, Xiao L, Zhang G, Xu R, Ji S, Li J (2017) A deep transfer learning approach for improved posttraumatic stress disorder diagnosis. In: IEEE international conference on data mining, pp 11–20
Baydogan MG (2015) Multivariate time series classification datasets. http://www.mustafabaydogan.com. Accessed 28 Feb 2019
Baydogan MG, Runger G, Tuv E (2013) A bagoffeatures framework to classify time series. IEEE Trans Pattern Anal Mach Intell 35(11):2796–2802
Bellman R (2010) Dynamic programming. Princeton University Press, Princeton
Benavoli A, Corani G, Mangili F (2016) Should we really use posthoc tests based on meanranks? Mach Learn Res 17(1):152–161
Bengio Y, Yao L, Alain G, Vincent P (2013) Generalized denoising autoencoders as generative models. In: International conference on neural information processing systems, pp 899–907
Bianchi FM, Scardapane S, Løkse S, Jenssen R (2018) Reservoir computing approaches for representation and classification of multivariate time series. arXiv:1803.07870
Bishop C (2006) Pattern recognition and machine learning. Springer, Berlin
Bostrom A, Bagnall A (2015) Binary shapelet transform for multiclass time series classification. In: Big data analytics and knowledge discovery, pp 257–269
Che Z, Cheng Y, Zhai S, Sun Z, Liu Y (2017a) Boosting deep learning risk prediction with generative adversarial networks for electronic health records. In: IEEE international conference on data mining, pp 787–792
Che Z, He X, Xu K, Liu Y (2017b) DECADE: a deep metric learning model for multivariate time series. In: KDD workshop on mining and learning from time series
Chen H, Tang F, Tino P, Yao X (2013) Modelbased kernel for efficient time series analysis. In: ACM SIGKDD international conference on knowledge discovery and data mining, pp 392–400
Chen H, Tang F, Tiño P, Cohn A, Yao X (2015a) Model metric colearning for time series classification. In: International joint conference on artificial intelligence, pp 3387–394
Chen Y, Keogh E, Hu B, Begum N, Bagnall A, Mueen A, Batista G (2015b) The UCR time series classification archive. www.cs.ucr.edu/~eamonn/time_series_data/. Accessed 28 Feb 2019
Chollet Fea (2015) Keras. https://keras.io. Accessed 28 Feb 2019
Chouikhi N, Ammar B, Alimi AM (2018) Genesis of basic and multilayer echo state network recurrent autoencoders for efficient data representations. arXiv:1804.08996
Cristian Borges Gamboa J (2017) Deep learning for timeseries analysis. arXiv:1701.01887
Cui Z, Chen W, Chen Y (2016) Multiscale convolutional neural networks for time series classification. arXiv:1603.06995
Dau HA, Silva DF, Petitjean F, Forestier G, Bagnall A, Keogh E (2017) Judicious setting of dynamic time warping’s window width allows more accurate classification of time series. In: IEEE international conference on big data, pp 917–922
Dau HA, Bagnall A, Kamgar K, Yeh CCM, Zhu Y, Gharghabi S, Ratanamahatana CA, Keogh E (2018) The UCR time series archive. ArXiv arXiv:1810.07758
Demšar J (2006) Statistical comparisons of classifiers over multiple data sets. Mach Learn Res 7:1–30
Deng H, Runger G, Tuv E, Vladimir M (2013) A time series forest for classification and feature extraction. Inf Sci 239:142–153
Esling P, Agon C (2012) Timeseries data mining. ACM Comput Surv 45(1):12:1–12:34
Faust O, Hagiwara Y, Hong TJ, Lih OS, Acharya UR (2018) Deep learning for healthcare applications based on physiological signals: a review. Comput Methods Programs Biomed 161:1–13
Forestier G, Petitjean F, Dau HA, Webb GI, Keogh E (2017) Generating synthetic time series to augment sparse datasets. In: IEEE international conference on data mining, pp 865–870
Friedman M (1940) A comparison of alternative tests of significance for the problem of \(m\) rankings. Ann Math Stat 11(1):86–92
Gallicchio C, Micheli A (2017) Deep echo state network (DeepESN): a brief survey. arXiv:1712.04323
Garcia S, Herrera F (2008) An extension on “statistical comparisons of classifiers over multiple data sets” for all pairwise comparisons. Mach Learn Res 9:2677–2694
Geng Y, Luo X (2018) Costsensitive convolution based neural networks for imbalanced timeseries classification. ArXiv arXiv:1801.04396
Glorot X, Bengio Y (2010) Understanding the difficulty of training deep feedforward neural networks. In: International conference on artificial intelligence and statistics, vol 9, pp 249–256
Goldberg Y (2016) A primer on neural network models for natural language processing. Artif Intell Res 57(1):345–420
Gong Z, Chen H, Yuan B, Yao X (2018) Multiobjective learning in the model space for time series classification. IEEE Trans Cybern 99:1–15
Grabocka J, Schilling N, Wistuba M, SchmidtThieme L (2014) Learning timeseries shapelets. In: ACM SIGKDD international conference on knowledge discovery and data mining, pp 392–401
Hatami N, Gavet Y, Debayle J (2017) Classification of timeseries images using deep convolutional neural networks. In: International conference on machine vision
He K, Zhang X, Ren S, Sun J (2015) Delving deep into rectifiers: surpassing humanlevel performance on imagenet classification. In: IEEE international conference on computer vision, pp 1026–1034
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: IEEE conference on computer vision and pattern recognition, pp 770–778
Hills J, Lines J, Baranauskas E, Mapp J, Bagnall A (2014) Classification of time series by shapelet transformation. Data Min Knowl Discov 28(4):851–881
Hinton G, Deng L, Yu D, Dahl GE, Mohamed AR, Jaitly N, Senior A, Vanhoucke V, Nguyen P, Sainath TN, Kingsbury B (2012) Deep neural networks for acoustic modeling in speech recognition: the shared views of four research groups. IEEE Signal Process Mag 29(6):82–97
Hoerl AE, Kennard RW (1970) Ridge regression: applications to nonorthogonal problems. Technometrics 12(1):69–82
Holm S (1979) A simple sequentially rejective multiple test procedure. Scand J Stat 6(2):65–70
Höppner F (2016) Improving time series similarity measures by integrating preprocessing steps. Data Min Knowl Discov 31:851–878
Hu Q, Zhang R, Zhou Y (2016) Transfer learning for shortterm wind speed prediction with deep neural networks. Renew Energy 85:83–95
Ignatov A (2018) Realtime human activity recognition from accelerometer data using convolutional neural networks. Appl Soft Comput 62:915–922
Ioffe S, Szegedy C (2015) Batch normalization: accelerating deep network training by reducing internal covariate shift. In: International conference on machine learning, vol 37, pp 448–456
Ismail Fawaz H, Forestier G, Weber J, Idoumghar L, Muller PA (2018a) Data augmentation using synthetic data for time series classification with deep residual networks. In: International workshop on advanced analytics and learning on temporal data, ECML PKDD
Ismail Fawaz H, Forestier G, Weber J, Idoumghar L, Muller PA (2018b) Evaluating surgical skills from kinematic data using convolutional neural networks. In: Medical image computing and computer assisted intervention, pp 214–221
Ismail Fawaz H, Forestier G, Weber J, Idoumghar L, Muller PA (2018c) Transfer learning for time series classification. In: IEEE international conference on big data, pp 1367–1376
Jaeger H, Haas H (2004) Harnessing nonlinearity: predicting chaotic systems and saving energy in wireless communication. Science 304(5667):78–80
Kate RJ (2016) Using dynamic time warping distances as features for improved time series classification. Data Min Knowl Discov 30(2):283–312
Keogh E, Mueen A (2017) Curse of dimensionality. In: Encyclopedia of Machine Learning and Data Mining. Springer, pp 314–315
Kim Y (2014) Convolutional neural networks for sentence classification. In: Empirical methods in natural language processing
Kingma DP, Ba J (2015) Adam: a method for stochastic optimization. In: International conference on learning representations
Kotsifakos A, Papapetrou P (2014) Modelbased time series classification. In: Advances in intelligent data analysis, pp 179–191
Krasin I, Duerig T, Alldrin N, Ferrari V, AbuElHaija S, Kuznetsova A, Rom H, Uijlings J, Popov S, Kamali S, Malloci M, PontTuset J, Veit A, Belongie S, Gomes V, Gupta A, Sun C, Chechik G, Cai D, Feng Z, Narayanan D, Murphy K (2017) OpenImages: a public dataset for largescale multilabel and multiclass image classification. https://storage.googleapis.com/openimages/web/index.html. Accessed 28 Feb 2019
Krizhevsky A, Sutskever I, Hinton GE (2012) ImageNet classification with deep convolutional neural networks. In: Advances in neural information processing systems vol 25, pp 1097–1105
Kruskal JB, Wish M (1978) Multidimensional scaling. Number 07–011 in sage university paper series on quantitative applications in the social sciences
Längkvist M, Karlsson L, Loutfi A (2014) A review of unsupervised feature learning and deep learning for timeseries modeling. Pattern Recognit Lett 42:11–24
Large J, Lines J, Bagnall A (2017) The heterogeneous ensembles of standard classification algorithms (HESCA): the whole is greater than the sum of its parts. arXiv:1710.09220
Le Q, Mikolov T (2014) Distributed representations of sentences and documents. In: International conference on machine learning, vol 32, pp II–1188–II–1196
Le Guennec A, Malinowski S, Tavenard R (2016) Data augmentation for time series classification using convolutional neural networks. In: ECML/PKDD workshop on advanced analytics and learning on temporal data
LeCun Y, Bottou L, Bengio Y, Haffner P (1998a) Gradientbased learning applied to document recognition. Proc IEEE 86(11):2278–2324
LeCun Y, Bottou L, Orr GB, Müller KR (1998b) Efficient backprop. In: Montavon G (ed) Neural networks: tricks of the trade. Springer, Berlin, pp 9–50
LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521:436–444
Lin S, Runger GC (2018) GCRNN: groupconstrained convolutional recurrent neural network. IEEE Trans Neural Netw Learn Syst 99:1–10
Lines J, Bagnall A (2015) Time series classification with ensembles of elastic distance measures. Data Min Knowl Discov 29(3):565–592
Lines J, Taylor S, Bagnall A (2016) HIVECOTE: the hierarchical vote collective of transformationbased ensembles for time series classification. In: IEEE international conference on data mining, pp 1041–1046
Lines J, Taylor S, Bagnall A (2018) Time series classification with HIVECOTE: the hierarchical vote collective of transformationbased ensembles. ACM Trans Knowl Discov Data 12(5):52:1–52:35
Liu C, Hsaio W, Tu Y (2018) Time series classification with multivariate convolutional neural network. IEEE Trans Ind Electron 66:1–1
Lu J, Young S, Arel I, Holleman J (2015) A 1 tops/w analog deep machinelearning engine with floatinggate storage in 0.13 \(\mu \)m cmos. IEEE J SolidState Circuits 50(1):270–281
Lucas B, Shifaz A, Pelletier C, O’Neill L, Zaidi N, Goethals B, Petitjean F, Webb GI (2018) Proximity forest: an effective and scalable distancebased classifier for time series. Data Min Knowl Discov 28:851–881
Ma Q, Shen L, Chen W, Wang J, Wei J, Yu Z (2016) Functional echo state network for time series classification. Inf Sci 373:1–20
Malhotra P, TV V, Vig L, Agarwal P, Shroff G (2018) TimeNet: pretrained deep recurrent neural network for time series classification. In: European symposium on artificial neural networks, computational intelligence and machine learning, pp 607–612
Martinez C, Perrin G, Ramasso E, Rombaut M (2018) A deep reinforcement learning approach for early classification of time series. In: European signal processing conference
Mehdiyev N, Lahann J, Emrich A, Enke D, Fettke P, Loos P (2017) Time series classification using deep learning for process planning: a case from the process industry. Proc Comput Sci 114:242–249
Mikolov T, Chen K, Corrado G, Dean J (2013a) Efficient estimation of word representations in vector space. In: International conference on learning representations—workshop
Mikolov T, Sutskever I, Chen K, Corrado G, Dean J (2013b) Distributed representations of words and phrases and their compositionality. In: Neural information processing systems, pp 3111–3119
Mittelman R (2015) Timeseries modeling with undecimated fully convolutional neural networks. arXiv:1508.00317
Neamtu R, Ahsan R, Rundensteiner EA, Sarkozy G, Keogh E, Dau HA, Nguyen C, Lovering C (2018) Generalized dynamic time warping: unleashing the warping power hidden in pointwise distances. In: IEEE international conference on data engineering
Nguyen TL, Gsponer S, Ifrim G (2017) Time series classification by sequence learning in allsubsequence space. In: IEEE international conference on data engineering, pp 947–958
Nwe TL, Dat TH, Ma B (2017) Convolutional neural network with multitask learning scheme for acoustic scene classification. In: AsiaPacific signal and information processing association annual summit and conference, pp 1347–1350
Nweke HF, Teh YW, Algaradi MA, Alo UR (2018) Deep learning algorithms for human activity recognition using mobile and wearable sensor networks: state of the art and research challenges. Expert Syst Appl 105:233–261
Ordón̈ez FJ, Roggen D (2016) Deep convolutional and LSTM recurrent neural networks for multimodal wearable activity recognition. Sensors 16:115
Pan SJ, Yang Q (2010) A survey on transfer learning. IEEE Trans Knowl Data Eng 22(10):1345–1359
Papernot N, McDaniel P (2018) Deep knearest neighbors: towards confident, interpretable and robust deep learning. arXiv:1803.04765
Pascanu R, Mikolov T, Bengio Y (2012) Understanding the exploding gradient problem. arXiv:1211.5063
Pascanu R, Mikolov T, Bengio Y (2013) On the difficulty of training recurrent neural networks. In: International conference on machine learning, vol 28, pp III–1310–III–1318
Petitjean F, Forestier G, Webb GI, Nicholson AE, Chen Y, Keogh E (2016) Faster and more accurate classification of time series by exploiting a novel dynamic time warping averaging algorithm. Knowl Inf Syst 47(1):1–26
Poggio T, Mhaskar H, Rosasco L, Miranda B, Liao Q (2017) Why and when can deepbut not shallownetworks avoid the curse of dimensionality: a review. Int J Autom Comput 14(5):503–519
Rajan D, Thiagarajan J (2018) A generative modeling approach to limited channel ecg classification. In: IEEE engineering in medicine and biology society, vol 2018, p 2571
Rajkomar A, Oren E, Chen K, Dai AM, Hajaj N, Liu PJ, Liu X, Sun M, Sundberg P, Yee H, Zhang K, Duggan GE, Flores G, Hardt M, Irvine J, Le Q, Litsch K, Marcus J, Mossin A, Tansuwan J, Wang D, Wexler J, Wilson J, Ludwig D, Volchenboum SL, Chou K, Pearson M, Madabushi S, Shah NH, Butte AJ, Howell M, Cui C, Corrado G, Dean J (2018) Scalable and accurate deep learning for electronic health records. NPJ Digit Med 1:18
Ratanamahatana CA, Keogh E (2005) Three myths about dynamic time warping data mining. In: SIAM international conference on data mining, pp 506–510
Rowe ACH, Abbott PC (1995) Daubechies wavelets and mathematica. Comput Phys 9(6):635–648
Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A, Bernstein M, Berg AC, FeiFei L (2015) ImageNet large scale visual recognition challenge. Int J Comput Vis 115(3):211–252
Sainath TN, Mohamed AR, Kingsbury B, Ramabhadran B (2013) Deep convolutional neural networks for LVCSR. In: IEEE international conference on acoustics, speech and signal processing, pp 8614–8618
Santos T, Kern R (2017) A literature survey of early time series classification and deep learning. In: International conference on knowledge technologies and datadriven business
Schäfer P (2015) The BOSS is concerned with time series classification in the presence of noise. Data Min Knowl Discov 29(6):1505–1530
Serrà J, Pascual S, Karatzoglou A (2018) Towards a universal neural network encoder for time series. Artif Intell Res Dev Curr Chall New Trends Appl 308:120
Silva DF, Giusti R, Keogh E, Batista G (2018) Speeding up similarity search under dynamic time warping by pruning unpromising alignments. Data Min Knowl Discov 32(4):988–1016
Srivastava N, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R (2014) Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res 15:1929–1958
Strodthoff N, Strodthoff C (2019) Detecting and interpreting myocardial infarction using fully convolutional neural networks. Physiol Meas 40(1):015001
Susto GA, Cenedese A, Terzi M (2018) Timeseries classification methods: review and applications to power systems data. In: Big data application in power systems, pp 179 – 220
Sutskever I, Vinyals O, Le QV (2014) Sequence to sequence learning with neural networks. In: Neural information processing systems, pp 3104–3112
Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A (2015) Going deeper with convolutions. In: IEEE conference on computer vision and pattern recognition, pp 1–9
Tanisaro P, Heidemann G (2016) Time series classification using time warping invariant echo state networks. In: IEEE international conference on machine learning and applications, pp 831–836
Tripathy R, Acharya UR (2018) Use of features from RRtime series and EEG signals for automated classification of sleep stages in deep neural network framework. Biocybern Biomed Eng 38:890–902
Uemura M, Tomikawa M, Miao T, Souzaki R, Ieiri S, Akahoshi T, Lefor AK, Hashizume M (2018) Feasibility of an AIbased measure of the hand motions of expert and novice surgeons. Comput Math Methods Med 2018:9873273
Ulyanov D, Vedaldi A, Lempitsky V (2016) Instance normalization: the missing ingredient for fast stylization. arXiv:1607.08022
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I (2017) Attention is all you need. In: Advances in neural information processing systems, pp 5998–6008
Wang Z, Oates T (2015a) Encoding time series as images for visual inspection and classification using tiled convolutional neural networks. In: Workshops at AAAI conference on artificial intelligence, pp 40–46
Wang Z, Oates T (2015b) Imaging timeseries to improve classification and imputation. In: International conference on artificial intelligence, pp 3939–3945
Wang Z, Oates T (2015c) Spatially encoding temporal correlations to classify temporal data using convolutional neural networks. arXiv:1509.07481
Wang L, Wang Z, Liu S (2016a) An effective multivariate time series classification approach using echo state network and adaptive differential evolution algorithm. Expert Syst Appl 43:237–249
Wang W, Chen C, Wang W, Rai P, Carin L (2016b) Earlinessaware deep convolutional networks for early time series classification. arXiv:1611.04578
Wang Z, Song W, Liu L, Zhang F, Xue J, Ye Y, Fan M, Xu M (2016c) Representation learning with deconvolution for multivariate time series classification and visualization. arXiv:1610.07258
Wang S, Hua G, Hao G, Xie C (2017a) A cycle deep belief network model for multivariate time series classification. Math Probl Eng 2017:1–7
Wang Z, Yan W, Oates T (2017b) Time series classification from scratch with deep neural networks: a strong baseline. In: International joint conference on neural networks, pp 1578–1585
Wang J, Chen Y, Hao S, Peng X, Hu L (2018a) Deep learning for sensorbased activity recognition: a survey. Pattern Recognit Lett
Wang J, Wang Z, Li J, Wu J (2018b) Multilevel wavelet decomposition network for interpretable time series analysis. In: Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery and data mining, pp 2437–2446
Wilcoxon F (1945) Individual comparisons by ranking methods. Biom Bull 1(6):80–83
Yang Q, Wu X (2006) 10 challenging problems in data mining research. Inf Technol Decis Mak 05(04):597–604
Ye L, Keogh E (2011) Time series shapelets: a novel technique that allows accurate, interpretable and fast classification. Data Min Knowl Discov 22(1):149–182
Yosinski J, Clune J, Bengio Y, Lipson H (2014) How transferable are features in deep neural networks? In: International conference on neural information processing systems, vol 2, pp 3320–3328
Zeiler MD (2012) ADADELTA: an adaptive learning rate method. arXiv:1212.5701
Zhang C, Bengio S, Hardt M, Recht B, Vinyals O (2017) Understanding deep learning requires rethinking generalization. In: International conference on learning representations
Zhao B, Lu H, Chen S, Liu J, Wu D (2017) Convolutional neural networks for time series classification. Syst Eng Electron 28(1):162–169
Zheng Y, Liu Q, Chen E, Ge Y, Zhao JL (2014) Time series classification using multichannels deep convolutional neural networks. In: Webage information management, pp 298–310
Zheng Y, Liu Q, Chen E, Ge Y, Zhao JL (2016) Exploiting multichannels deep convolutional neural networks for multivariate time series classification. Front Comput Sci 10(1):96–112
Zhou B, Khosla A, Lapedriza A, Oliva A, Torralba A (2016) Learning deep features for discriminative localization. In: IEEE conference on computer vision and pattern recognition, pp 2921–2929
Ziat A, Delasalles E, Denoyer L, Gallinari P (2017) Spatiotemporal neural networks for spacetime series forecasting and relations discovery. In: IEEE international conference on data mining, pp 705–714
Acknowledgements
The authors would like to thank the creators and providers of the datasets: Hoang Anh Dau, Anthony Bagnall, Kaveh Kamgar, ChinChia Michael Yeh, Yan Zhu, Shaghayegh Gharghabi, Chotirat Ann Ratanamahatana, Eamonn Keogh and Mustafa Baydogan. The authors would also like to thank NVIDIA Corporation for the GPU Grant and the Mésocentre of Strasbourg for providing access to the cluster. The authors would also like to thank François Petitjean and Charlotte Pelletier for the fruitful discussions, their feedback and comments while writing this paper. This work was supported by the ANR TIMES project (Grant ANR17CE230015) of the French Agence Nationale de la Recherche.
Author information
Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Responsible editor: Dr. Eamonn Keogh.
Rights and permissions
About this article
Cite this article
Ismail Fawaz, H., Forestier, G., Weber, J. et al. Deep learning for time series classification: a review. Data Min Knowl Disc 33, 917–963 (2019). https://doi.org/10.1007/s10618019006191
Received:
Accepted:
Published:
Issue Date:
Keywords
 Deep learning
 Time series
 Classification
 Review