Keywords

1 Introduction

Sciences of various domains other than computer science use scientific workflows to model their complex computational pipelines, which brings them many benefits such as reusing the results of parts or entire workflows, failure management, or parallelisation. Managing the workflows’ execution is a complex task, as each workflow requires different computing, memory or I/O capacity, making the designing of a common, but appropriate, distributed environment for all workflows types very difficult and sometimes almost impossible. Workflows can also be executed in Cloud as similar as in traditional clusters, as many workflow management services that allow the effective utilisation of the Cloud’s elastic resources already exist [23]. Still, Cloud produces many additional challenges compared with the traditional clusters [2] caused by its on-demand elastic resource provisioning, dynamic starting of instances [20], and variant performance of virtual machines (VMs) during a time period [22].

Instead of executing in a real Cloud environment to determine the behavior of an application, many researchers resort to simulators for their analysis [3], which allows them to reduce costs for purchasing and maintaining expensive hardware resources, and time for executing time-consuming algorithms [21] and later on to determine that its performance is over- or under estimated. Moreover, simulators can be used to experiment new prototype solutions and identify the “optimal” resource configurations before deployment of production platforms. Most common simulators allow users to create a virtual data center considering their latest computing, networking, energy, or cost requirements. However, although they can simulate an elastic Cloud data center, simulators usually neglect the Cloud performance fluctuation and uncertainty [10], which can lead to wrong estimation. This aspect is especially important for workflow executions, as they consist of a high number of data and control flow dependencies [15] that further affect their overall performance without any correlation [18]. A small disturbance in a task execution can dramatically affect the scheduling of the following workflow tasks, resulting in a completely inefficient schedule that increases the execution cost or execution time (the makespan), or sometimes even both [1].

Although some simulators, such as DynamicCloudSim [4], allow users to configure the simulation considering a certain heterogeneity and instability of Cloud, still, there are several deficiencies for a proper configuration. The configuration itself is a complex process, as users need to configure more than ten parameters, for which they do not know the exact values of parameters in order to configure a specific Cloud and application to be executed. For example, the default values that are intended for Amazon’s EC2 are several years old, and cannot be used for other public or private Clouds. Additionally, public Clouds can be seen as a black-box whose internal parameters are unknown to regular users. Even by using results of previous research, some parameters still cannot be generalized as they are valid either for that specific Cloud, a specific application, or even a combination of both.

We therefore propose a simpler model that reduces the configuration of the noisiness to a single parameter the Cloud noisiness, instead of dozens. The injected noise causes an instability in Task Execution Time (TET), which improves the accuracy, represented through the trueness (i.e. closeness of the true mean value) and the precision (i.e. closeness of corresponding standard deviation) of the simulation, as defined in ISO-5725 standard [12]. We conduct a series of evaluation experiments in Amazon EC2 and the most common state-of-the-art simulator - DynamicCloudSim. The evaluation proves that our simpler model implemented in DynamicCloudSim shows up to \(9.2\%\) higher trueness and up to 8.39 times higher precision for workflow execution simulations compared to DynamicCloudSim. Not only that our proposed model shows better accuracy to the real execution, but its configuration is much simpler and easier.

The paper is structured in several sections as follows. Section 2 presents the related works in modeling the workflow execution instability and the features of cloud simulators in this domain. The models for workflow, Cloud, experiments and test cases that are used for our noisiness model are presented in Sect. 3. Our simplified, more accurate model of adding a noise in simulation is described in Sect. 4. Sections 5 and 6 present the testing methodology and results of the evaluation of our model and current state-of-the-art simulation model of DynamicCloudSim. Finally, we conclude the paper and present plans for future work in Sect. 7.

2 Related Work

Many Cloud features and parameters can cause the performance instability: heterogeneity of resources, instance types, number of instances, instance straggling, instance failures, multi-tenancy, networking bottlenecks, resource time-sharing, etc. As a consequence, an instance of the same type provides different performance for the same task over some time period. Dejun et al. [7] reported high performance deviations in Amazon EC2. Jackson et al. [13] determined that different underlying hardware for similar instances caused performance perturbation. Schad et al. [19] detected a long-term performance instability of Amazon EC2, which was correlated also to the CPU model of the same instance type, the hour of the day, and the day of the week. Iosup et al. [11] determined yearly and daily patterns of performance variability, but also periods of constant performance. All these behaviors depend also on the executed application.

CloudSim [5] simulates scheduling algorithms and resource provisioning in elastic Cloud environment, but Cloud performance instability remains unaddressed. Chen and Deelman [6] extended the Cloudsim into WorkflowSim, by introducing several parameters specific to workflows. Still, all these extensions do not introduce the Cloud performance instability. Other works developed scalable simulators covering up to hundreds of thousands of heterogeneous machines [8]. For example, GroudSim [16] is a scalable event-based simulator for Grid and Cloud environments. GloudSim [9] is a simulator that introduces some dynamics in execution by resizing the instances. Still, it does not offer a TET’s instability, as the performance of specific VM is constant during a time period.

Bux and Leser [4] went further in this direction by developing the DynamicCloudSim simulator as an extension of CloudSim that introduces several additional characteristics to simulate the Cloud heterogeneity, such as heterogeneous underline hardware, VM stragglers, VM failures, long and short term fluctuations, etc. However, configuring dozens of parameters for heterogeneity is not an easy task, as users are usually not aware of the internal Cloud architecture. We therefore went a step further by treating the Cloud as a black-box and introduced much simpler approach that needs a configuration of the noise into one parameter only. It includes two instabilities in itself: workflow noisiness (e.g. dependencies, structure, TET deviation) and Cloud noisiness (e.g. heterogeneity). Nevertheless, although it is a simple method, the results of the evaluation show that our model improves the accuracy compared to the related DynamicCloudSim’s instability model. Schad et al. [19] reported that several performance parameters are unstable with a normal distribution. We also use the normal distribution to add a noise in TET, but instead of naively generating the variables distributed with a normal distribution, we inject the noisiness parameter by shifting the TET’s mean value by the Cloud noisiness parameter.

3 Modeling the Workflow and Cloud

This section formally models the workflow and Cloud environment, which are used for our cloud noisiness model later on.

3.1 Workflow Application Model

We model a workflow application W as a precedence constraint graph (TD) consisting of a set \(T=\bigcup _{i=1}^{n}{\{T_{i}\}}\) of n tasks \(T_i\), which are interconnected through a set of dependencies \(D=\left\{ \left( T_i, T_{j}, D_{ij}\right) |\left( T_i, T_j\right) \in T\times T\right\} \), where \(\left( T_i, T_j, D_{ij}\right) \) implies that \(T_i\) needs to be executed before \(T_j\), and the file size to be transferred from \(T_i\) to \(T_j\) is \(D_{ij}\) bytes. The tasks are assumed to be non-preemptive, so it is not allowed to suspend one and resume it later on.

The function \( pred : T \rightarrow \mathcal {P}(T)\), where \(\mathcal {P}\) denotes the power set, returns the set of immediate predecessors of each task \(T_i \in T\) (i.e. \(T_j \in pred \left( T_i\right) \iff \left( T_j, T_i, D_{ji}\right) \in D\)), while the function \( succ : T \rightarrow \mathcal {P}(T)\) returns the set of immediate successors of the task \(T_i\) (i.e. \(T_j \in succ \left( T_i\right) \) \( \iff \left( T_i, T_j, D_{ij}\right) \in D\)). Each workflow has an entry task \(T_{ entry }\) with no predecessors (i.e. \(T_{ entry } \in T : pred \left( T_{ entry }\right) \) \( = \emptyset \)) and an exit task \(T_{ exit }\) with no successors (i.e. \(T_{ exit } \in T : succ \left( T_{ exit }\right) = \emptyset \)).

Each task \(T_i\) has a requirement vector \(R_i\), which defines its hardware or software requirements such as the minimum value of memory or storage needed for execution. We express the computational complexity \(w_i\) (i.e. work) of each task \(T_i\) in million of instructions (MI).

Note that this is a simplified view of a workflow. Still, workflows with multiple entry/exit tasks are covered by adding a single “dummy” entry task with the computational complexity \(w=0\) (i.e. without any complexity) before all entry tasks or a single “dummy” exit task, also with the computational complexity \(w=0\), after all exit tasks, respectively.

3.2 Cloud Infrastructure Model

A Cloud offers a set of r VM types \(IT=\bigcup _{k=1}^{r}{\left\{ IT_k \right\} }\). Each instance type \(IT_k\) is characterized by two parameters: computational speed \(s_k\) in million instructions per second (MIPS) and number of CPUs \(c_k\). We denote the set of available VM instances as: \(I=\bigcup _{j=1}^{m}{\left\{ I_j \right\} }\), whose number m is constant during the workflow execution. Each instance \(I_j\) has an associated instance type \(IT_k\) defined as a function: \( type : I \rightarrow IT\).

We model the expected TET \(t_i^j\) of a task \(T_i\) as the ratio between its computational complexity and the speed of the instance \(I_j\) on which it is executed (\(t_i^j = w_i / s_k\), \(I_j \in I \wedge IT_k = type \left( I_j\right) \)).

The completion time of a task \(T_i\) executed on an instance \(I_j\) is the latest completion time of all its predecessors plus its expected TET:

$$\begin{aligned} end \left( T_i{,k} \right) =\left\{ \begin{matrix} t_{i}^j, &{} T_i = T_{ entry }{, IT_k = type(I_j)};\\ \underset{T_p \in pred \left( T_i\right) }{\max }{\left\{ end \left( T_p{,k}\right) +t_{i}^j \right\} }, &{} T_i \ne T_{ entry }{, IT_k = type(I_j)}. \end{matrix}\right. \end{aligned}$$

In a single experiment, we are using a constant number of the same type VMs, which keep running. Thus, for an experiment that uses VM instances of type k, we define the workflow makespan \(M=end(T_{exit}{,k})\).

3.3 Experiment and Test Case Model

In order to model the environment-independent model, we define a set of experiments \(EXP = \bigcup _{x=1}^{q}{\left\{ EXP_x \right\} }\). Each element \(EXP_x\) is modeled as a triple \(EXP_x(W, IT_{k}, {v})\), which means that a workflow W is executed on a specific number of VMs v, all of the same type \(IT_k\).

As we want to simulate Cloud’s behavior, we repeat each experiment N times and we refer to each execution as a test case. Therefore, a test case \(^xTC^c\) represents the c-th repetition of an experiment \(EXP_x\). This means that an experiment can be considered as a matrix of N columns (workflow execution repetitions) and n rows (tasks within a workflow).

Let \(^xt_i^c\) denotes the measured TET of a task \(T_i\) in a test case \(^xTC^c\) of an experiment \(EXP_x\). Analogue, the makespan of test case \(^xTC^c\) will be denoted as \(^xM^c\). As we want to analyze the distribution of makespans per experiment, we define the TET’s mean value \(^x\overline{t}_i = \frac{1}{N} \cdot \sum _{c=1}^{N} {^xt_i^c}\) of a task \(T_i\) and the mean makespan \(^x\overline{M} = \frac{1}{N} \cdot \sum _{c=1}^{N} {^xM^c}\). Both mean values are defined for an experiment \(EXP_x\).

4 Noise Simulation Model

In this section we present our new model of simulating a workflow execution in Cloud, which improves the accuracy through a much simpler approach that uses only one parameter for noising for all the tasks of a workflow - Cloud noisiness, instead of DynamicCloudSim’s 13 parameters for heterogeneity and instability.

4.1 Workflow Noisiness

In order to define a model for the noisiness provided by the workflow, first we model the noisiness of a task \(T_{i}\in T\) itself with the parameter TET’s deviation \(^x\rho _{i}^{cd}\), which is defined in (1) as a relative TET difference of task \(T_{i}\) in two test cases \(^xTC^{c}\) and \(^xTC^{d}\) of an experiment \(EXP_x\) with \(c,d \in [1,N] \wedge c \ne d\).

$$\begin{aligned} ^x\rho _{i}^{cd} = \frac{|^{x}t_{i}^{c}-{^{x}t_{i}^{d}}|}{max(^xt_{i}^c,^xt_i^d)} \end{aligned}$$
(1)

The TET’s deviation \(^x\rho _{i}^{cd}\) is used to introduce the workflow noisiness \(^x\overline{\varDelta }^{cd}\), which describes the total noisiness of all TETs within a workflow in two test cases \(^xTC^{c}\) and \(^xTC^{d}\) of an experiment \(EXP_x\), as defined in (2). Formally, it represents a normalized mean TET’s difference of all corresponding tasks \(T_{i}\) of the same workflow, in two test cases \(^xTC^c\) and \(^xTC^d\) of an experiment \(EXP_x\).

$$\begin{aligned} {\begin{matrix} ^x\overline{\varDelta }^{cd} = \frac{1}{n} \cdot \sum _{i=1}^{n} {^x\rho _{i}^{cd}} \end{matrix}} \end{aligned}$$
(2)

The workflow noisiness \(^x\overline{\varDelta }^{cd}\) can be used to extract noisiness of Cloud environment and includes vertical average of TETs of all tasks within a workflow W. \(^x\overline{\varDelta }^{cd}\) is an intermediate metric and serves as input for the Cloud noisiness, which is explained in the following subsection.

4.2 Cloud Noisiness

The workflow noisiness shows the instability of a workflow in two executions only. As Cloud environment is unstable, we want to take some average of a set of test cases for a single experiment \(EXP_x\). Therefore, we introduce the Cloud noisiness \(^x\overline{\varDelta }\) of an experiment \(EXP_x\). As defined in (3), \(^x\overline{\varDelta }\) represents the average experiments’ makespan instability of all N repetitions (test cases) of an experiment \(EXP_x\).

$$\begin{aligned} {\begin{matrix} ^x\overline{\varDelta } = \frac{1}{\frac{N \cdot (N - 1)}{2}} \cdot \sum _{\forall c,d | 1 \le c < d \le N}{^x\overline{\varDelta }^{cd}} \end{matrix}} \end{aligned}$$
(3)

We measure the workflow noisiness \(^x\overline{\varDelta }^{cd}\) of each unique pair of test cases \(^xTC^c\) and \(^xTC^d\) for all N executions (test cases) of the same experiment \(EXP_x\). Accordingly, the total number of unique pairs that can be generated from a set of N elements is \(\left( {\begin{array}{c}N\\ 2\end{array}}\right) = \frac{N \cdot (N- 1)}{2}\). The Cloud noisiness \(^x\overline{\varDelta }\) includes horizontal average of all workflow executions (test cases) within a single experiment \(EXP_x\).

4.3 Modeling the Noising

Now, after formally definition of cloud noisiness, we define how to model and add a noise in a simulation. The noised TET \(^x\widetilde{\tau }_i\) of a task \(T_i\) is defined in (4), where Gaussian(MeanSTDEV) represents a random function with Gaussian (normal) distribution.

$$\begin{aligned} {\begin{matrix} ^x\widetilde{\tau }_i = (1 + ^x\overline{\varDelta } + Gaussian(0, \sigma _{^x\overline{\varDelta }})) \cdot ^x\overline{t}_i \end{matrix}} \end{aligned}$$
(4)

Accordingly, the noise in (4) is modeled as a Gaussian distribution, where the mean value is the TET’s mean value \(^x\overline{t}_{i}\) of a task \(T_i\), shifted with the Cloud noisiness \(^x\overline{\varDelta }\) in order to cover uncertain overheads, measured in the experiments. For noisiness, we use the standard deviation \(\sigma _{^x\overline{\varDelta }}\) of all workflow noisiness of a single cloud noisiness, which is determined by each pair of test cases \(TC^c\) and \(TC^d\) (\(\forall c,d~|~1 \le c < d \le N\)) of a single experiment.

5 Testing Methodology

In this section we present the testing methodology in order to evaluate our noising model in Sect. 6.

5.1 Synthetic Workflow

The synthetic workflow that is used in our experiments consists of two parallel sections (Second and Fourth) of same size, with three synchronisation tasks (First, Third and Fifth) in between, as depicted in Fig. 1. The chosen workflow structure is the result of workflow characteristics analysis of several well known workflows, such as EPIGENOMICS and SIPHT [14]. The workflow size is related to the number of tasks in the parallel sections. In the experiments we use a parallel section size of 13 (SYNWF/13) and 44 (SYNWF/44). We use two different workflow sizes in order to cover balanced and unbalanced executions, such that one is a prime number and the other is dividable by the numbers 2 and 4, corresponding to the number of instances. Additionally, we used instance number of 3, which is not a divisor of neither workflow sizes. With this workflow structure and selected parameters, we also want to investigate if there is a correlation between the workflow parameters and the execution environment (chosen resources and the inefficient workflow execution).

Fig. 1.
figure 1

The structure of the synthetic workflow with the file transfer to computation time ratios

Table 1. File transfer (FT) to computation time (CT) ratio in % for SYNWF/13 and SYNWF/44

As workflow makespan consists of computations and file transfers (including both the network and I/O), we have chosen different file transfer to computation time ratios for all five task types, as depicted in Fig. 1 and presented in Table 1. A file transfer describes the copying process of a set of files from one task to another and combines network bandwidth with I/O. The ratios of the First, Second and Fourth task types are assumed to be constant for different workflow sizes because the number and size of file inputs is not changing. This is different for the Third and Fifth task types, which are synchronization tasks and collect all the output files produced by the Second and Fourth tasks (parallel sections), correspondingly. Thus, in order to compensate this, we slightly increase the file transfer time for the workflow size 44. Moreover, the ratios of both parallel sections are inverse to each other.

5.2 Cloud Testing Environment

All Cloud experiments were executed in Amazon EC2 with the VM image Amazon Linux AMI (ami-1ecae776) in the availability zone US East (N. Virginia). For the workflow execution and measurements of the TETs and makespan, we used the workflow execution engine Askalon [17]. We use two Amazon instance types \(IT=\{t2.small, t2.medium\}\) as well as the number of instances \(m=\{2,3,4\}\).

According the definition of experiments in Sect. 3.3, \(EXP_1\) is defined as \(EXP_{{1}}(SYNWF/13, t2.small, 2)\), which means that it executes the synthetic workflow with parallel section size 13, by using two small VMs. As we used two different workflows that are specified in the previous subsection, we define \(q = 12\) experiments, and since we execute 20 repetitions (test cases) of each experiment, we execute a total of 240 test cases in Cloud. In order to cover different behavior of Amazon’s EC2, we run all test cases of each experiments in the period of two weeks.

5.3 Simulation Testing Environment

In all our simulations, we used DynamicCloudSim, which extends CloudSim by adding features that allow a user to simulate the heterogeneity in Cloud described as the performance deviation of a resource, including VM heterogeneity, host heterogeneity, File I/O heterogeneity, and Cloud instability, VM Stragglers and VM Failures. DynamicCloudSim introduces 13 parameters that cover Cloud heterogeneity and instability, which are described in Table 2.

Table 2. DynamicCloudSim heterogeneity parameter setup for \(S_{sDCS}\) and \(S_{noise}\)

The same two workflows, two VM types and two, three, and four number of VMs are used to reproduce the same 12 experiments as in real Cloud. We execute \(N = 20\) test cases per experiment in order to have equal number of experiments in real and simulated executions. Note that we consider the instance startup as warm up period and thus it is discarded in the experiments.

In order to evaluate our model, we compare it with the DynamicCloudSim’s model (denoted as \(S_{sDCS}\)) with default values for Amazon EC2 (Table 2), which are based on other performance-based researches, experience, and assumptions.

Our noisining model \(S_{noise}\), adds noise to one parameter (TET) only. Thus, all other heterogeneity and noise related parameters are set to 0 in DynamicCloudSim. Note that also the cpuNoiseCV parameter is also set to 0, because we insert our noise through the Cloud noisiness parameter. Table 3 shows the measured Cloud noisiness \(^x\overline{\varDelta }\) and the corresponding standard deviations for \(S_{noise}\) in each experiment executed on Amazon EC2. It shows that EC2 provides computation instabilities from \(8.9\%\) up to \(17.6\%\) for various workflows and using different number of various instance types. The deviation of all test cases per a single experiment is in the range of 2.8% up to 11.2%.

Table 3. \(^x\overline{\varDelta }\) and \(\sigma _{^x\overline{\varDelta }}\) values for \(S_{noise}\)

For a fair comparison, both simulations have equal base network, storage and computation speed. Moreover during the experiments in real Cloud, we did not detect any VM failures, and therefore the straggler, failure and cpu/io/bw BaselineChangesPerHour parameters are set to 0 for both simulations.

6 Evaluation

In this section we present the results of a series of experiments to evaluate our noising model. The summary of the evaluation shows not only that our model is simpler, but it is more accurate than DynamicCloudSim.

Figure 2 shows the mean makespans of both workflows SYNWF/13 and SYNWF/44. The experiments are denoted as the product of number of instances (m) and the abbreviation of VM type (M for t2.medium and S for t2.small VM type). For example, \(2*S\) denotes the experiment that uses two small instances, while \(4*M\) is used for the experiment with four medium instances.

Fig. 2.
figure 2

Mean makespans with variations \(\pm \sigma (mean~makespan)\) of SYNWF/13 and SYNWF/44 using two, three or four instances of type t2.small (S) or t2.medium (M)

The mean values of makespans of all experiments with SYNWF/13 are depicted in Fig. 2a, along with the standard deviation. We observe that our \(S_{noise}\) model shows higher accuracy, that is, both higher trueness and precision, than \(S_{sDCS}\). Increasing the number of small instances, \(S_{noise}\) improves its precision with \(20.16\%\) up to \(71.61\%\), while \(S_{sDCS}\) has minimum \(51.21\%\) of precision offset compared to C. For all experiments, our noise model shows higher makespan trueness for real Cloud C, compared to \(S_{sDCS}\). In detail, the trueness’ offset of \(S_{noise}\) is between \(6.0\%\) and \(10.7\%\) for small instances, while \(S_{sDCS}\)’s is between \(11.8\%\) and \(13.0\%\). Thus, \(S_{noise}\) is up to \(7.0\%\) better than \(S_{sDCS}\) for small instances. We observe a similar behavior with increasing number of medium instances, where the makespan trueness offset of \(S_{noise}\) is between \(0.5\%\) and \(14.7\%\), while \(S_{sDCS}\)’s is between \(2.3\%\) and \(16.3\%\). In the experiments with two medium instances, both simulations show higher makespan than C, but \(S_{noise}\) has still higher trueness with \(0.5\%\) offset.

Figure 2b shows the average makespan results for all experiments with the SYNWF/44 workflow. Similar to the experiments with the SYNWF/13 workflow, our model \(S_{noise}\) shows again higher accuracy than \(S_{sDCS}\). The trueness of our model \(S_{noise}\) is better than \(S_{sDCS}\) for all experiments, while the precision is comparable, but slightly worse, only for experiments with 4 instances, and for all others our model \(S_{noise}\) is still better. In detail, \(S_{noise}\) shows between \(15.29\%\) and \(20.59\%\) higher standard deviation than C for 2 and 3 instances, while \(S_{sDCS}\) has up to 5 times higher offset. On experiments with 4 instances, \(S_{noise}\) shows between \(76.94\%\) and \(86.21\%\) precision offset, while \(S_{sDCS}\) has only up to \(21.93\%\). Regarding the simulated makespans’ trueness (closeness to the mean value), \(S_{noise}\) shows a better trueness than \(S_{sDCS}\) in all experiments. In detail, comparing the trueness offset of the simulated results and the real Cloud results, \(S_{noise}\) is still better with trueness offset between \(9.4\%\) and \(26.6\%\), while \(S_{sDCS}\) has between \(16.0\%\) and \(28.8\%\).

Comparing all experiments conducted with both workflows, \(S_{noise}\) shows higher and more closer precision than \(S_{sDCS}\) for all experiments with SYNWF/13. We observe similar behavior with SYNWF/44, except for experiments with 4 small and medium instances. Additionally, \(S_{noise}\) shows up to \(9.2\%\) higher makespans’ trueness compared to \(S_{sDCS}\) for all experiments. We also observe that the precision of both models does not depend if the number of instances is a divisor or not of the workflow parallel section size, which can significantly reduce the number of experiments to determine the cloud noisiness.

7 Conclusion and Future Work

This paper presents a new simplified Cloud noisiness model for noising the workflow execution while it is simulated in order to behave as the real Cloud unstable environment. Instead of configuring dozens parameters in order to achieve a noised simulation, as it is required in DynamicCloudSim, our model configures only one - the Cloud noisiness. A series of experiments in Amazon EC2 that were reproduced in simulated environment show that our Cloud noisiness model simplifies the simulation configuration and improves the simulation trueness, and especially precision.

The main novelty in our model is the calculation of noisiness. Instead of using the normal distribution naively for the tasks’ runtime, we shift the mean TET by the Cloud noisiness and then add a noise (deviation of the Cloud noisiness) of each task. The workflow noisiness smooths the impact of the workflow structure by calculating the average instability of all tasks in a workflow, while the Cloud noisiness estimates the environmental noise by calculating the horizontal average instability of the same task. With these two parameters, we inject not only the noise of a task itself, when being executed in Cloud, but we inject the impact of common tasks’ noises within the workflow and environmental noise provided when a whole workflow is being repeatedly executed in that environment.

Although our Cloud noisiness approach requires several executions of a workflow in order to calculate the Cloud noisiness parameter, the results show that the instability is not correlated to the parallel section size, that is, if we execute a workflow efficiently or inefficiently. However, the experiments show that the instability is highly correlated with the instance type and the number of instances. We will try to extend our model with other parameters in order to reduce this dependency and therefore the cost of learning the Cloud noisiness. However, our analysis shows that the performance instability of up to 17% between two experiments is comparable with the standard deviation of up to 11% between test cases within a single experiment.

Our noisiness model improved the simulator’s trueness for all and the precision for most instance types and number of instances. As the results show that the instability is instance type dependent, we will extend our model towards modeling network, as well as I/O, and including them in order to improve the trueness and precision even more.