Introduction

Statistical process control (SPC) is a tremendous quality assurance tool to develop the quality of manufacture and ultimately scores on end-customer satisfaction. SPC uses control charts to monitor the most important key quality characteristics (KQCs) in manufacturing (Sharma and Rao 2013).

In general, there is at least a small variation on quality characteristics of the produced items. Hence, we should control the processes to reduce the amount of nonconforming products. Control charts basically are used to monitor processes to become aware on any alteration that may affect the quality of product.

Generally, SPC control charts are used to detect changes in a process by distinguishing between assignable causes and common causes of the process variation. When a control chart signals, process engineers initiate a search to identify and eliminate the source of variation. Knowing the time at which the process began to vary, the so-called change point would help to conduct the search more efficiently in a tighter time-frame (Assareh et al. 2013; Akhavan Niaki and Khedmati 2013). Control charts also are used to detect anomalies in the processes. They are most often used to monitor production-related processes. In many business-related processes, the quality of a process or product can be characterized by a relationship between a response variable and one or more explanatory variables which is referred to as profile (Narvand et al. 2013; Soleimani et al. 2013).

In many applications, quality of process is characterized by a single random variable called quality characteristic but some cases occur that process is characterized by more than one quality characteristic. These random variables are usually correlated and jointly distributed and cannot be controlled independently using a univariate control chart. Accordingly, multivariate statistical control methods have been proposed to investigate this issue. Most of the works on control charts in multivariate case are on problem of monitoring mean vector of the process. A measure of distance that takes into account the covariance structure was proposed by Harold Hotelling (1931). It is called Hotelling’s T 2 in honour of its developer. Geometrically we can view T2 as proportional to the squared distance of a multivariate observation from the target where equidistant points form ellipsoids surrounding the target. The higher the T2 value, the more distance the observation from the target is. In the Hotelling’s T2 control chart the mean vector and covariance matrix are unknown and must be estimated by means of the previous data where it may affect the performance of control chart. Recently, some researchers such as Tchao and Hawkins (2011), Capizzi and Masarotto (2010) and Jensen et al. (2006) proposed solutions to investigate this issue.

When a control chart is used to monitor a process, three design parameters that should be selected are the sample size, the sampling interval, and the action limit(s). Duncan (1956) offered an economic model incorporated the most important relevant cost items associated with sampling and control charts. Through minimization of the proposed cost model, the optimum economical design parameters of control chart were presented.

In the literature, statistical control chart design may be applied to increase the power of any control chart such as T2. Aparisi (1996) followed this idea through adaptive sample size and sampling interval in the multivariate case and proposed three types of modified charts with variable sample size (VSS), variable sampling interval (VSI), and variable sample size and sampling interval (VSSI) features, respectively (see Aparisi 1996; Aparisi and Haro 2001, 2003), given that the mean vector and variance–covariance matrix were known. Chen and Hsieh (2007) indicated that traditional T 2 chart gives a better performance if both sample size and control limits are variable (VSSC), and the waiting time between successive samples are fixed.

In the case of economic design of control charts, Chen (2009, 2007) used a Markov chain approach to design VSI T 2 and VSSI T 2 control charts. He showed that both of them can be more efficient than FSR (Fixed Sampling Rate) control scheme in terms of the loss. Chou et al. (2006) developed a cost function for variable sampling intervals T2 control charts and obtained optimum design parameters using genetic algorithms. Costa and Rahim (2001) used the Markov chain approach to reach an economic design of X ¯ charts with variable parameters. De Magalhaes et al. (2001) developed a cost model for economic design of X ¯ chart with all design parameters varying in an adaptive way. They check whether the economic model for a Vp X ¯ chart reduces the quality cost of a process. Bai and Lee (1998) presented an economic design of the VSI X ¯ control charts and showed that the VSI scheme can be more efficient than the FSI scheme in terms of the expected cost per time. They applied a two-stage optimization approach to find the optimal sampling-and-charting parameters of their cost model.

In this paper, we propose a novel economic design of T2 control chart based on the extension of Duncan’s (1956) cost model. By using a genetic algorithm, the optimal design parameters of the relevant cost model besides the sensitivity analysis is proposed through an illustrative example.

The rest of the paper is organized as follows: in “Variable sample size and control limits T2 control chart” we briefly review the VSSC T2 control chart. We then present a method and describe the proposed formulation for the of cost model for multivariate situations in “Proposed cost model”. “Illustrated example” systematically guides readers to implement the proposed procedure via a numerical example. Finally, we close with a conclusion.

Variable sample size and control limits T2 control chart

Let X ¯ 1 , X ¯ 2 , X ¯ 3 , be p × 1 random vectors, each representing sample mean vector of related quality characteristics assumed jointly distributed as p-variate normal with mean vector μ0 and variance–covariance matrix 0 When ith sample of size n is taken at every sampling point, we calculate the following statistic:

χ i 2 = n · X ¯ i - μ 0 · Σ 0 - 1 · X ¯ i - μ 0 i = 1 , 2 , 3
(1)

and compare it with upper control limit (or action limit) denotes by UCL χ 2 which can be specified by the ( 1 - α ) percentile point of a χ2 distribution with p degree of freedom χ p , α 2 . However, in most cases the values of μ 0 and 0 are unknown and are estimated by sample mean vector ( X ¯ ¯ ), and sample variance–covariance matrix (S) of m initial random samples prior to on-line process monitoring and T2 statistic is defined by

T i 2 = n · X ¯ i - X ¯ ¯ · S - 1 · X ¯ i - X ¯ ¯ ; i = 1 , 2 , m
(2)

that is the approximate statistic for Hotelling’s multivariate chart. In this case, action limit used to monitor future random vectors is given by Alt (1984) as

k = C ( m , n , p ) · F p , v , α .
(3)

where F p , v , α is the ( 1 - α ) percentile point of F distribution with p and v degrees of freedom. C ( m , n , p ) and v are calculated by

C ( m , n , p ) = p · ( m + 1 ) · ( n - 1 ) m n - m - p + 1 ; n > 1 p · ( m + 1 ) · ( m - 1 ) m 2 - m p ; n = 1 , v = m n - m - p + 1 ; n > 1 m - p ; n = 1
(4)

Traditional Hotelling’s T2 chart operates with a fixed sample of size n 0 drawn every h 0 hours from process, and T2 statistic is plotted on a control chart with k 0 = C ( m , n 0 , p ) · F p , v , α as the action limit. The VSSC T2 chart is a modification of traditional T2 chart. Let n 1 , w 1 , k 1 be minimum sample size, largest warning and action limits, and n 2 , w 2 , k 2 be maximum sample size, smallest warning and action limits, respectively, such that n 1 < n 0 < n 2 while keeping sampling interval fixed at h. The warning ( w j ) and action ( k j = C ( m , n j , p ) · F p , v , α ) limits divide T2 chart to three regions as shown in Table 1:

Table 1 Three regions in VSSC T 2 chart

The decision to switch between maximum and minimum sample size depends on position of the prior sample point on the control chart and summarizes as following function:

n ( i ) , w ( i ) , k ( i ) = n 1 , w 1 , k 1 n 2 , w 2 , k 2 if if 0 T i - 1 2 w i - 1 w i - 1 < T i - 1 2 k i - 1
(5)

During the in-control period, it is assumed that the size of samples is chosen at random between two values when the process starts or after a false alarm. Small size is selected with probability of p 0 , whereas large sample size is selected with probability of 1 - p 0 , where p 0 is the conditional probability of a sample point falling in the safe region, given that it did not fall in the action region and is calculated as follows:

p 0 = Pr T i 2 < w 1 | T i 2 < k 1 = Pr T 1 2 < w 2 | T i 2 < k 2
(6)
1 - p 0 = Pr w 1 < T i 2 < k 1 | T i 2 < k 1 = Pr w 2 < T i 2 < k 2 | T i 2 < k 2
(7)

Proposed cost model

Cost model is an extension of Duncan (1956) model which was employed in a univariate case. First, we make a number of assumptions as follows:

  • The mean vector and variance–covariance matrix of process are unknown.

  • At beginning, the process is in-control but after a random time it will be disturbed by an assignable cause that causes a fixed shift in the process mean vector.

  • The process after the shift remains out of control until the assignable cause is eliminated (if possible).

  • When the T i 2 value falls outside the action limit, the process is stopped and then a search is started to find the assignable cause and adjust the process.

  • The interval between starting the process and occurring of an assignable cause follows an exponential distribution with λ as its parameter.

In the economic design of VSSC T2 control chart we tend to find the optimal design parameters that minimize the expected cost per time. Figure 1 depicts the production cycle, which is divided into four time intervals of in-control period, out-of-control period, searching period due to false alarm, and the time period for identifying and correcting the assignable cause. Individuals are now illustrated before they are grouped together.

Fig. 1
figure 1

Production cycle considered in the cost model

( T 1 ) The expected length of in-control period is 1 λ .

( T 3 ) The expected length of out-of-control state represents the average time needed for the control chart to produce a signal after the process mean shift. T 3 is given by

T 3 = h · ARL out - t + n · G
(8)

where G is the average time from taking a sample to the time of plotting T i 2 statistic on the chart, and n is the average sample size when the process operates in out-of-control state, and ARL out is the average number of samples drawn from process when it is out of control. n and ARL out are given by

n = p 0 · n 1 · E ( U 1 ) + n 2 · E N i 1 E ( N 1 ) + 1 - p 0 · n 1 · E N i 2 + n 2 · E ( U 2 ) E ( N 2 )
(9)
ARL out = p 0 · E ( N 1 ) + 1 - p 0 · E ( N 2 )
(10)

where as indicated by Chen (2007a, b), E U 1 is the average number of sample points in the safe region when the process is in out-of-control state and current sample point belongs to safe region. Then,

E U 1 = 1 - p 22 1 + p 11 · p 22 - p 11 - p 22 - p 12 · p 21
(11)

E U 2 is the average number of sample points in the warning region when the process is in out-of-control state and current sample point belongs to warning region. Then,

E ( U 2 ) = 1 - p 11 1 + p 11 · p 22 - p 11 - p 22 - p 12 · p 21
(12)

E N i 1 is the average number of sample points in the warning region when the process is in out-of-control state and current sample point belongs to safe region. Then,

E N i 1 = 1 + p 12 1 - p 22
(13)

E(N 2 i ) is the average number of sample points in the safe region when the process is in out-of-control state and current sample point belongs to warning region. Then,

E N i 2 = 1 + p 21 1 - p 11
(14)

E ( N 1 ) is the average number of samples drawn from the time of process mean shift to the time that mean shift is detected given that first sample point after mean shift falls into the safe region. Then,

E ( N 1 ) = 1 - p 22 + p 12 1 + p 11 · p 22 - p 11 - p 22 - p 12 · p 21
(15)

E ( N 2 ) is the average number of samples drawn from the time of the process mean shift to the time that mean shift is detected given that first sample point after mean shift falls into the warning region. Then,

E ( N 2 ) = 1 - p 11 + p 21 1 + p 11 · p 22 - p 11 - p 22 - p 12 · p 21
(16)

where

p 11 = Pr T i 2 < w 1 | T i 2 C ( m , n 1 , p ) · F p , v 1 , τ 1
p 12 = Pr w 1 < T i 2 < k 1 | T i 2 C ( m , n 1 , p ) · F p , v 1 , τ 1
p 21 = Pr T i 2 < w 2 | T i 2 C ( m , n 2 , p ) · F p , v 2 , τ 2
p 22 = Pr w 2 < T i 2 < k 2 | T i 2 C ( m , n 2 , p ) · F p , v 2 , τ 2

where C m , n j , p and v j for j = 1 , 2 is calculated by Eq. (4), and F p , v j , τ j for j = 1 , 2 is a random variable followed non-central F distribution with p and v j degrees of freedom and non-centrality parameter τ j defined by τ j = n j · μ 1 - μ 0 - 1 μ 1 - μ 0 . If we define δ = μ 1 - μ 0 - 1 μ 1 - μ 0 , then τ j = n j · δ 2 , where δ is the Mahalanobis distance that is a measure of change in process mean vector.

( T 2 ) Let t 0 denote the average amount of time exhausted searching for the assignable cause when the process is in-control, and E ( F A ) denote the expected number of false alarms per cycle, which is given by

E ( FA ) = 1 h λ ARL i n = 1 h λ 1 α = α · 1 h λ ;
(17)

then, the expected length of searching period due to false alarms can be expressed by T 2 = t 0 · E ( FA ) .

( T 4 ) The time to identify and correct the assignable cause following an action signal is a constant t 1 .

Aggregating the foregoing four time intervals, the expected length of a production cycle would be expressed by

E ( T ) = 1 λ + t 0 · α · 1 h λ + h · ARL out - t + n · G + t 1
(18)

If one defines C 0 , the average search cost if the given signal is false; C 1 , the average cost to discover the assignable cause and adjust the process to in-control state; C 2 , the hourly cost when the process is operating in control state; C 3 , the hourly cost when the process is operating in out-of-control state; C 4 , the cost for each inspected item; then the expected cost during a production cycle is given by

E ( C ) = C 0 · E ( FA ) + C 1 + C 2 · 1 λ + C 3 · h · ARL out - t + n · G + C 4 · E ( N )
(19)

where E ( N ) is the average number of inspected items and is calculated by

E ( N ) = n · E ( N in ) + n · ARL out
(20)

where given that the process is in-control, E ( N in ) is the average numbers of samples drawn from the process, and n is the average sample size. They are given by

E ( N i n ) = 1 / λ h = 1 h λ
(21)
n = n 1 · p 0 + n 2 · 1 - p 0
(22)

Finally, the expected cost per time ECT is given by

ECT = E ( C ) E ( T )
(23)

Illustrated example

The usefulness and effectiveness of the proposed procedure beside the optimal approximation and sensitivity analysis on main parameters is demonstrated using a numerical example which acts as a modification of Lin et al. (2009). Suppose that a production process is monitored by the VSSC T2 control chart. The cost and process parameters are as shown in Table 2,

Table 2 Cost and process parameters for numerical example

The cost model given in Eq. (23) has some specification abbreviated as follows:

  • It is a nonlinear model and a function of mixed continuous-discrete decision variable

  • Mathematically, model space is a discrete and non-convex.

Hence, using nonlinear programming techniques for optimizing this model is time consuming and inefficient. Hence, we decided to use the Genetic Algorithms (GA) introduced by Holland (1975) with a mathematical software package (MATLAB 7.1) to obtain the optimal values of n 1 , n 2 , h , w 1 , w 2 , k 1 , k 2 that minimize the expected cost per time. Some advantages of GA are as follows:

  • GA uses the fitness function and the stochastic concepts (not deterministic rule) to search for optimal solution. Therefore, the GA can be applied for many kinds of optimization problems.

  • Mutation and crossover techniques in the GA avoid trapping in the local optimum.

  • The GA is able to search for many possible solutions at the same time.

  • We applied the solution procedure used in Lin et al. (2009) to our example as follows:

Step 1. Initialization Generating 30 initial solutions randomly, which satisfy the following constraints:

n 1 < n 2 , w 1 < w 2 , k j = C ( m , n j , p ) · F p , v j , α , 0 < h < 20

Step 2. Evaluation Calculating the value of the cost function in Eq. (23) to evaluate each solution.

Step 3. Selection Replacing the solution with highest cost by the solution with lowest cost.

Step 4. Crossover Selecting a pairs of solutions in step 3 randomly to use them as the parents for crossover operations. In this example, we apply the arithmetical crossover method with crossover probability 0.3 as follows:

Offspring 1 = 0.3 Parents 1 + 0.7 Parent 2; Offspring 2 = 0.7 Parents 1 + 0.3 Parents 2

Where offsprings are new chromosomes. At the end of GA steps, determination of crossover probability is described in detailed.

Step 5. Mutation Here, we use non-uniform method to carry out the mutation operation with the rate of 0.07. Thus, we can randomly select 7 % of chromosomes to mutate some parameters (or genes). At the end of GA steps, determination of mutation rate is described in detailed.

Step 6. Repeat Step 2–5 until the stopping criteria is found. In this example, we use “50 generations” as our stopping criteria.

For implementing the GA, we need to determine its parameters: the population size (PS), the crossover Probability (CP), the mutation rate (MR), and the number of generations (GN). Here, we use the orthogonal array experiment to determine the values of these parameters. As shown in Table 3, three levels of each parameter are planned in this orthogonal array experiment. An L9 orthogonal array is employed and the GA parameters are then assigned to it. In the L9 orthogonal array experiment, there are nine different level combinations of the four parameters. For each trial or combination, three cost values, denoted by y 1 , y 2 , and y 3 , are obtained from the GA and the results are recorded in Table 4.

Table 3 Level plan for the three control parameters in the GA
Table 4 Experimental layout of L9 array and the experimental results

Based on the information in Table 5, the optimal level combination of the four control parameters in the GA is that PS = 30, CP = 0.3, MR = 0.07, and GN = 50.

Table 5 Response table of S/N’s for the three control parameters in the GA

By running MATLAB for different values of process mean shift, we achieved the optimal approximate solution of the example as shown in Tables 6 and 7.

Table 6 Solutions of the cost model for different process mean shifts in VSSC scheme
Table 7 Solution of the cost model for different process mean shifts in FSR scheme

In order to investigate the effect of model’s parameters on the final solution, sensitivity analysis is arranged using orthogonal-array experimental design and multiple linear regression analysis.

Based on the proposed model, n 1 , n 2 , h , w 1 , w 2 , k 1 , k 2 are determined as the responses. Eleven control factors (C 0 ,C 1 ,C 2 ,C 3 ,C 4 ,m,… t 1 ) each with three levels, shown in Table 8, are allocated sequentially to an L33 orthogonal array, as shown in Table 9. The experiments are conducted randomly. The experimental data were analysed by following the proposed procedure strictly. For each trial, genetic algorithm was applied to produce the best approximate solution of the economic design of VSSC T2 chart and the results are recorded in Table 10.

Table 8 Different levels of model and cost parameters
Table 9 Experimental design based on the L 33 orthogonal array
Table 10 The optimal approximate solution of the proposed cost model of the VSSCT2 control chart

Consecutively to examine the effects of parameters on the responses, regression analysis concerned by Minitab statistical package. The outputs of Minitab include the ANOVA and regression coefficients tables beside normal probability plot of residuals evaluated for models adequacy and validity which show the final set of regression lines and summary of regression models (Table 11) as follows:

Table 11 Summary of regression models and ANOVA table
n 1 = - 0.885401 + 0.06375 C 3 + 0.0277627 m + 2.6875 δ
n 2 = 41.6714 + 0.0975 C 3 - 0.7625 C 4 - 18.75 δ - 10.3125 t 0
h = 10.7 + 0.183 C 2 - 0.0713 C 3 + 4.23 δ - 81.5 λ
w 1 = 1.69 + 0.755 C 4 - 0.0771 m + 3.42 p - 8.23 δ
w 2 = 0.231 + 0.640 p
k 1 = 11 + 1.60 C 4 - 0.22 m + 7.92 p - 17.4 δ
k 2 = 7.54 + 0.0272 m + 2.36 p + 0.785 δ
ECT = 9.23 + 0.506 C 2 + 0.355 C 3 + 0.894 C 4 - 20.5 δ + 448 λ
AATS = 268 + 9.70 C 2 - 4.21 C 3 + 20.6 C 4 - 223 δ
E ( FA ) = 0.0059 + 0.000844 C 3 + 0.000486 m - 1.65 λ
ARL out = 16.1 + 1.24 C 4 - 18.3 δ

The first estimated regression line indicates the hourly cost when the process operates in out-of-control state ( C 3 ), the number of rational subgroups (m) and the magnitude of mean shift (δ) affect the small sample size ( n 1 ). The second estimated line shows the Minitab output for the large sample size. Seeing the regression line, the hourly cost when the process operates in out-of-control state (C3), the cost of inspecting each item ( C 4 ), the magnitude of mean shift (δ) and the average time wasted due to searching for assignable cause when the process is in control ( t 0 ), significantly influence the value of large sample size ( n 2 ). The sign of coefficient of C 3 is positive indicating a larger magnitude of C 3 results in a larger amount of n 2 . Also, Because of the coefficients of C 4 , δ and t 0 are negative, by increasing each of them the value of large sample size decreases. The estimated small action limit regression line ( k 2 ) noted that a higher number of rational subgroups (m) will reduce the amount of k 2 . On the other hand, if the number of quality characteristics (p) and the magnitude of mean shift (δ) increases, the amount of k 2 increases. Regression line of optimal value of cost function (ECT) is affected significantly by three cost parameters and two process parameters (i.e., C 2 , C 3 , C 4 , λ , δ ). A larger shift magnitude of process mean (δ) leads a lower value of ECT. Meanwhile, increase in values of C2, C3, C4 and λ results in increase in the value of ECT. Also similar analysis may be conducted for the sampling interval (h), the large warning limit (w1), the small warning limit (w2), the large action limit (k1), adjusted average time to signal (AATS), the average number of false alarms (E(FA)), and the number of samples drawn when the process operates in out-of-control state ( ARL out ), respectively.

Concluding remarks

Delivering economical design of the VSSC T 2 control chart on the presence of fixed sampling intervals and exponentially distributed assignable causes is the main contribution of the present study which provides more sensitivity in the traditional Hotelling’s T2 control chart in rapid detecting of small drifts in the process mean vector. The real assumption on the occurrence times of the assignable cause is allowed us in applying the Markov chain approach on constructing the proposed expected hourly cost model as a novel extension of the priors. The main accomplished results on the proposed model are

  • Larger changes in the process mean vector cause to increase value of small action limit. Additionally, it tends to generate a lower expected cost per time and large sample size.

  • The large sample size tends to be raised when the hourly cost of operating process in out-of-control state increases. Also, it decreases when cost of inspecting each item or wasted time due to each false alarm increases.

  • By growth in value of the hourly cost of operating the process in control, the hourly cost of operating the process out of control or the cost for each inspected item, the expected cost per time increases.

  • The small action limit will be large by adding to the number of quality characteristics or deduction in the number of rational subgroups.

  • If the duration of in control period increases, the expected cost per time will decrease.