1 Introduction

In job shop scheduling problem (JSSP), the objective is to allocate resources in a way, such that a number of tasks can be completed cost-effectively within a given set of constraints. JSSP is one of the most widely studied problems in computer science, which has great importance to manufacturing industries with the objective to minimise the production cost. Classical JSSP can be described as scheduling n different jobs on m machines. The m machines are identical, and the n nonpreemptable jobs are all independent. Job i has the size J i  > 0, 1 ≤ i ≤ n, which is to be processed on a set of m machines (M r ), 1 ≤ r ≤ m. Each job has a technological sequence of machines to be processed. The jobs arrive one by one, and each job must be immediately and irrevocably scheduled without knowledge of later jobs. The size of a job is known on arrival, and the jobs are executed only after the scheduling is completed. The processing of job J i on machine M r is called the operation O ir. Operation O ir requires the exclusive use of M r for an uninterrupted duration P ir , its processing time. A schedule is a set of completion time for each operation C ir that satisfies these constraints. Thus, JSSP can be considered as a searching or optimization problem, where the goal is to find the best possible schedule. The classical JSSP is one of the most challenging combinatorial optimization problems, mainly because of two reasons. Firstly, even for the special case of m = 2, JSSP is NP-hard. Secondly, one has to deal with tricky search space as neighboring points may represent very different solutions.

Due to expansion in manufacturing industry and industrial automation, JSSP has become a practical problem in the industry and started getting a lot of attention from the research community. The difficulty of the general JSSP is that it makes rather hard for conventional search-based methods to find a set of optimal schedules within polynomial time. Deterministic scheduling methods like the branch-and-bound method (Bucker et al. 1994) or the dynamic programming (Peter et al. 1999) are always computationally expensive for an optimum solution when searching a large search space. To combat the increasing complexity, it has been suggested by many researchers for near-optimal solutions instead. Such near-optimal solutions can only be endowed by stochastic search techniques such as evolutionary algorithms (EAs) (Ripon 2007). Davis proposed the first GA-based approach to the solution of scheduling problems in 1985 (Davis 1985). This paper was instructive to the application of GA on JSSP. Since then, GA has been applied with increasing frequency to JSSP. In contrast to smart heuristics such as simulated annealing (Kirkpatrick et al. 1985) and tabu-search (Glover 1989) which are local search techniques and use a generate-and-test search manipulating one feasible solution based on physical rather than a biological analogy, the GA utilizes a population of solutions in its search, giving it more resistance to premature convergence on local minima.

Surprisingly most of the researches in JSSP have focused mainly on a single objective, and predominantly optimization of the makespan, which is defined as the time interval between start of first operation and completion of last operation. However, real world scheduling problems naturally involve multiple objectives. In general, minimization of the total makespan is often used as the optimization criterion in JSSP. However, tardiness, flow time, lateness, earliness are also the important criteria in JSSP. Inherently, real-life scheduling problems are multi-objective by nature and the final schedule must consider various objectives simultaneously. Consequently, scheduling falls into the category of multi-objective optimization problem (MOOP). In such MOOPs, there is no single optimal solution, rather there is a set of alternative solutions. These solutions, namely Pareto-optimal solutions (Deb 2001; Zitzler et al. 2000; Veldhuizen and Lamont 2000), are optimal in the wider sense that no other solutions in the search space are superior when all the objectives are considered. In a Pareto-optimal set, any two solutions of this set do not dominate each other and one or more members of this set dominate all other solutions in the search space excluding this set. Based on the principle of multi-objective optimization, obtaining an optimal solution that satisfies all objectives is almost impossible. It is mainly for the conflicting nature of objectives, where improving one objective may only be achieved when worsening another objective. Accordingly, it is desirable to obtain as many different Pareto-optimal solutions as possible, which should be converged to, and diverse along the Pareto-optimal front with respect to multiple criteria.

Nagar et al. (1995) provided a good review on multi-objective production scheduling. This survey mostly covered researches conducted up to early 1990s based on a complete classification scheme. In those researches, the solution methodology primarily comprises of implicit enumeration techniques such as branch and bound, dynamic programming and trade-off curves. They specifically mentioned that they did not come across any paper that used techniques like simulated annealing, tabu-search and genetic algorithm. This has also been confirmed in a survey by Lei (2009) that conventional techniques are the main approaches to multi- and bi-objective scheduling problems before 1995. Lei provided extensive survey coverage on the literature of multi-objective production scheduling, where he first classified scheduling problems into deterministic and non-deterministic classes. The deterministic scheduling problems are categorized into four types. The author surveyed more than 90 papers published between 1995 and 2008, out of which about 40 papers adopted evolutionary algorithm (EA) and GA as optimization methods. It is evident from the two surveys that only in the current decade a great deal of attention has been focused on solving JSSP using EAs (Ripon 2007; Bierwirth et al. 1996; Jensen 2003; Syswerda 1991; Bierwirth 1995; Yamada and Nakano 1992; Song et al. 1999; Chan et al. 2008; Ombuki and Ventresca 2004).

The fundamental concept of JSSP can be thought of as an ordering problem. A schedule is a representational issue which resembles to a permutation problem. A permutation problem can generally be formulated in the following way. A set of n operations (tasks) with known processing times has to be scheduled on m machines (resources). A group of m operations forms a complex called a job. Altogether N jobs are defined within the set of operations, i.e. the number of operations is the Cartesian product of the number of jobs and machines defined by O = N × m, where O is the number of operations. Partitioned permutation provides the bedrock for the application of EA to many combinatorial optimisation problems (Bierwirth et al. 1996) and thus serves as chromosomes in GA. The known snare of this permutation chromosome is the crossover operator, which may not safeguard the semantical properties of the underlying problem. Therefore, one of the central issues in the use of GAs for the JSSP is an efficient characterization of the strength of crossover operator. This is perhaps due to crossover operator’s most exploratory power in an GA. Eshelman et al. investigated the exploratory power of crossover operator (Eshelman et al. 1989). Bierwirth developed repeating permutation representation while Mattfeld implemented a number of crossover operations (Bierwirth et al. 1996; Bierwirth 1995; Mattfeld 1996). A large number of crossover operators have been proposed in the literature such as generalized partially mapped crossover (GPMX) (Bierwirth et al. 1996), generalized order crossover (GOX) (Bierwirth 1995), precedence preservation crossover (PPX) (Bierwirth et al. 1996; Mattfeld 1996) and such others, mainly due to the need for designing specialist crossover operations to use with permutation representations. The details of crossover operators specifically designed for ordering applications can be found in (Bierwirth et al. 1996; Jensen 2003). In all of these permutation based crossover techniques, there exist redundancies at the tail of a chromosome. A schedule, decoded from a chromosome sequence of redundant genes, has no effect on the final solution. It also imposes additional time complexity. Therefore, the objective of the current paper is to improve the crossover technique and propose the improved precedence preservation crossover (IPPX) to be used in the JSSP, which eventually resolves this issue and also reduces the execution time. In an attempt to address multiple objectives simultaneously in this work, we apply makespan and mean-flow time as the objectives and present the schedules as a set of Pareto-optimal solutions.

The remainder of the paper is organized as follows. Section 2 describes various crossover operators. Section 3 presents the proposed approach for the improvement of PPX. Experimental results on the performance of the proposed method (IPPX) and comparisons with others approaches are demonstrated in Sect. 4. Finally, some concluding remarks on the method are made in Sect. 5.

2 Crossover in permutation representation

In crossover, also known as recombination, information is exchanged among the chromosomes (or individuals) present in the mating pool to create new chromosomes. Permutation-based representations present particular difficulties for the design of crossover operators, since it is not generally possible simply to exchange strings of genes between selected parents and still maintain the permutation properties. The crossover operator has to comply with the semantic meaning (properties) of the chromosome representation meaning to combine building blocks to form larger building blocks which share the phenotypical traits of the smaller building blocks. Two genes may articulate meaningful information if they appear side by side (relative order). They may even articulate information if one gene precedes the other gene in the chromosome (absolute order) regardless of how many genes lay between. Syswerda (1991) inferred that the order as well as the position of genes in the permutation is meaningful. An excellent theoretical analysis on order-based crossover has been provided by Wroblewski (1996). To be more precise, we expect the absolute order to be of particular interest because it directly expresses precedence relation among the operations in a schedule. A number of specialized crossover operators have been designed for permutations. These aim at exchanging information as much as possible which especially held in common in both parents.

GPMX (Bierwirth et al. 1996) and GOX (Bierwirth 1995) assemble one offspring from two parent chromosomes (donator and receiver). In both techniques a substring is chosen from the donating chromosome. Then all genes of the substring are deleted with respect to their index of occurrence in the receiving chromosome. The GOX operator was first presented by Bierwirth (1995). GOX implants the substring into the receiver at the position where the first gene of the substring has occurred (before deletion) in the receiver. Unlike GOX and GPMX implants the substring at the position where it occurs in the donator. Figure 1 illustrates these two different crossover techniques.

Fig. 1
figure 1

GPMX in comparison to GOX for a typical 3 × 3 JSSP

PPX (Bierwirth et al. 1996; Jensen 2003) perfectly respects the absolute order of genes in parental chromosomes. First the offspring chromosome is initialized empty. Then a vector of length n is randomly filled with elements of the set {0, 1}. This vector defines the order in which genes are drawn from parent 1 and parent 2, respectively. After a gene is drawn from one parent and deleted from the other one, it is appended to the offspring chromosome. This step is repeated until both parent chromosomes become empty and the offspring contains all genes involved. Figure 2 depicts a typical PPX operation.

Fig. 2
figure 2

PPX for a typical 3 × 3 JSSP

In order to apply PPX in a uniform crossover fashion the choices may alternatively change at random. However, the absolute order between any of two genes in the offspring has its origin in at least one of the parental chromosomes. In short, GOX passes on the relative order of genes, GPMX tends to pass on positions of genes by respecting the ordering to some extent, and PPX respects the absolute order of genes resulting in a perfect preservation of precedence relations among genes.

2.1 Superiority of PPX over GPMX and GOX

Bierwirth (1995) discussed about the heredity of various schedule characteristics. He emphasized on precedence relations among operations in order to explain the representational issue of absolute order for the JSSP. In this approach, they map the permutation representation into a string of 0/1-decision expressing the absolute order of any of two operations to be processed on the same machine. This mapping leads to the binary representation proposed by Yamada and Nakano (1992). The binary mapping is very useful to measure the differences and similarities of schedules on a phenotypical level by means of the hamming distance. Crossover performed by GPMX, GOX or PPX results in one offspring. Ideally the offspring inherits one half of the characteristics from each parent. Table 1 describes the phenotypical distance d(p 1, p 2) between two randomly generated parents p 1 and p 2. The distances of the offspring, denoted as d(o, p i ) to both parents for 1,000 calls to PPX, GPMX and GOX are also shown in the table.

Table 1 Phenotypical preservation of crossover

The average normalized Hamming distance between two arbitrary solutions is 0.273. For all three crossover operators, we observe that d(o, p 1) ≈ d(o, p 2). This verifies all operators to pass on the same portion of parental information. In case of PPX, d(o, p 1) + d(o, p 2) = d(p 1, p 2) holds. It means PPX passes on precedence relations perfectly in comparison to that of GOX and GPMX.

2.2 Redundancy problem of PPX

Most of the GA approaches for JSSP represent a solution by a chromosome containing the sequence of all the operations and decode the chromosome to a real schedule from the first gene to the last gene. According to Song et al. (1999), there are three common problems for these approaches. Firstly, it introduces high redundancy at the tail of the chromosome. Secondly, there exists little significance of rear genes on the overall schedule quality. And finally, GA operators applied on the real part of the chromosome are less likely to create genetically improved good offspring, i.e., most likely a waste of evolution time.

It implies that the rear genes of a chromosome are not significant enough and consequently have less impact on overall schedule. The execution time of the last operation of a job does not reduce the waiting time of other operations of that job. As a result, it is nothing but waste of evolution time in case of applying the GA operators (crossover, mutation) on these genes. As for example, let us consider the previous 3 × 3 JSSP (Fig. 2) in the following Fig. 3.

Fig. 3
figure 3

Redundancies at the tail of the chromosome

Pointing at the last three genes (3, 2, and 1), we can find that all of these operations are the last operation of respective job. Consequently job 3, in the last four bits, we find the last operation for job 3. Now finishing the last operation of job 3 at an earlier time does not reduce the total completion time for job 3. So, it is redundant to manipulate these last genes. Most importantly, in case of large number of jobs and operations (Suppose 100 × 100 JSSP), this redundancy increases and makes unnecessary delay to deliver an optimal schedule.

3 Proposed improved PPX (IPPX) approach

3.1 Chromosome representation

Chromosome representations for JSSP using GA can be performed by two basic encoding approaches: direct and indirect (Chan et al. 2008). The direct approach encodes a schedule as a chromosome and the genetic operators are used to evolve these chromosomes into better schedules. In indirect representation, the chromosome encodes a sequence of decision preferences, like simple ordering of jobs in a machine or any heuristic rules, and a schedule builder is required to decode the chromosome into a schedule.

In this work, we applied indirect representation of chromosomes incorporated with a schedule builder. This is because, applying simple genetic operators on direct representation string often results in infeasible schedule solutions. Indirect chromosome representation is implemented with an un-partitioned operation-based representation where each job integer is repeated m times (m is the number of machines), and it is known as “permutation with repetition” (Bierwirth 1995) in mathematics. By scanning the permutation from left to right, the k-th occurrence of a job number refers to the k-th operation in the technological sequence of this job as depicted in Fig. 4. In this representation, any individual can be decoded into a feasible schedule, but two or more different individuals may be translated into an identical schedule. The advantage of such a scheme is that it requires a very simple schedule builder because all the generated schedules are legal.

Fig. 4
figure 4

Permutation with repetition approach for a 3 × 3 JSSP

3.2 Reduction of redundancy

As discussed earlier in Sect. 2, the permutation based crossover techniques exhibit redundancy at the tail of chromosomes. To reduce such tail redundancy, we perform PPX for (NM) genes from the two parent chromosome and choose the last M genes by using a simple heuristic method, where M follows the constraint \( O \le M \le \frac{N}{3}.\) Here, O is the number of operation and N is the total number of genes.

Suppose that P1 and P2 are the two chromosome (3 × 3 JSSP) using this method as shown in Fig. 5. Here, last (M = 3) genes are not taken under crossover operation. Instead, we choose the remaining M genes using the heuristic method. In this figure, the genes goes through crossover are shown by shadow and deleted genes in both chromosomes are represented by bold numbers. In order to adjust the total length of the chromosome, we apply a heuristic method for the rest of the genes in the offspring chromosome. In this approach, we always expect the last operations of a job is in the last M bits. From the two parent chromosomes, we choose that chromosome in which there is more number of last operations before the occurrence of a non-last operation in the remaining M bits. If this value is the same for both the parent chromosomes, then genes are taken from that parent who has a better fitness value. For example, for a 5 × 3 JSSP and M = 4, the last M genes of the chromosome are shown in Fig. 6.

Fig. 5
figure 5

Crossover for (N–M) genes

Fig. 6
figure 6

Selection of the last M genes by heuristics method

Here, we select the last four genes of the offspring from P1 as it contains more number of last operations before the occurrence of a non-last operation. If we observe closely, we see that in P0 there are two operations before the occurrence of a non-last operation (marked by a circle) whereas in P1 there are three operations before the occurrence of a non-last operation. As we always keep track of the operation number of a job, this method does not require additional efforts which makes it very simple, effective and less complex. Algorithm 1 presents the pseudo code for the proposed IPPX.

figure a

3.3 Mutation

We use swapping mutation technique in our approach. Here any two genes within (N–M) genes of the offspring chromosome are swapped. And the last M genes are left as they are, as shown in Fig. 7.

Fig. 7
figure 7

Mutation

3.4 Efficiency of IPPX method

Figure 8 represents the flowchart of the proposed GA based method for JSSP using IPPX. In our approach, we apply the GA operators on less number of genes in a chromosome. As the number of operation required for these operators are less than the usual ones, it saves execution time. In a generation, if the crossover rate is 0.8 and population size is 100, the number of crossover will be 80. So, for 1,000 generations there will be 80,000 crossovers. As the time required to complete a crossover is reduced, the total execution time will be reduced significantly. Thus, this approach is very effective when the population, generations, jobs and operations are huge in number. Real world scheduling problems are generally too large to find their optimal schedule in a reasonable amount of time. This is simply because the real world JSSPs include a larger number of jobs and machines as well as additional constraints and flexibilities; all of these in turn increase the complexity to a further level. So, we often look for near optimal solutions for a problem in a reasonable amount of time. The experimental results justify that our approach is able to achieve satisfactory and sometimes better than the average results in comparison with the existing methods.

Fig. 8
figure 8

Flowchart of the proposed method

4 Experimental result and analysis

4.1 Experimental setup

The experiments are conducted using 100 chromosomes and 150 generations. The probabilities of crossover and mutation are 0.9 and 0.3, respectively. Using the same setting, each problem is tested for 30 times with different seeds, and the best and average solutions are recorded. Then all the final generations were combined and a non-dominated sorting is performed to constitute the final non-dominated solutions. To justify the efficiency of the proposed crossover method (IPPX) (as shown in Algorithm 1) the results are compared to existing well-known crossover methods GPMX, GOX, and PPX in the framework for a multi-objective genetic algorithm. In this work, we use the non-dominated sorting genetic algorithm (NSGAII) (Deb et al. 2002) as the multi-objective GA framework. Additionally, to justify the efficiency of the proposed IPPX-based JSSP approach, we present a comparison with other GA-based JSSP approaches in single objective context. In addition, we compare the performance of the IPPX-based approach with the performance of other existing GA-based JSSP approaches in multi-objective context.

4.2 Benchmark problems

To evaluate how IPPX performs with respect to solution quality, we run the algorithm on various benchmark data. The first three well-known benchmark problems, known as mt06, mt10 and mt20, formulated by Muth and Thompson (1963) are commonly used as test beds to measure the effectiveness of a certain method. The mt10 and mt20 problems have been a good computational challenge for a long time. However, it is no longer a computational challenge now. Indeed, the mt10 problem has been referred as “notorious”, because it remained unsolved for over 20 years. The mt20 problem has also been considered as quite difficult. Applegate and Cook (1991) propose a set of benchmark problems called the “ten tough problems” as a more difficult computational challenge than the mt10 problem, by collecting difficult problems from literature, some of which still remain unsolved (Applegate and Cook 1991). Among these, the la problems are parts of 40 problems la01–la40 originated from (Lawrence 1984). The problem data and the lower bounds information are taken from the OR-library. Table 2 presents the problem size and the best known lower bounds for makespan for the problems that have been used in this study.

Table 2 Benchmark problems with their lower bounds

4.3 Objective functions

The quality criteria for a good schedule are: maximum tardiness, average tardiness, weighted flow time, weighted lateness, weighted tardiness, weighted number of tardy jobs and weighted earliness plus weighted tardiness (Fang et al. 1996). In the present work, the makespan has been considered as the first objective. The mean flow time, as the second objective, continues to be very important since it assists to select the appropriate one when many algorithms proposed in the literature have reached the same makespan for many instances. The two objective functions makespan and mean flow-time are defined by Eqs. 1 and 2, respectively. Note that, both of these objectives are subject to minimization.

$$ {\text{Makespan}} = { \min }\left\{ {{ \max }\left\lfloor C \right\rfloor } \right\} $$
(1)
$$ {\text{Mean flow-time}} = \frac{1}{n}\sum\limits_{i = 1}^{n} {C_{i} } $$
(2)

where C i is the completion time of job i and n is the number of jobs.

4.4 Analysis of results

4.4.1 Single objective context

Until now, almost all JSSP algorithms try to optimize single criteria only (mainly minimization of the makespan). Therefore, to evaluate our proposed IPPX based algorithm as an evolutionary approach, we first compared the makespan obtained by our approach with the existing GA-based approaches to justify its capability to optimize makespan. We then demonstrated its performance as a multi-objective evolutionary JSSP algorithm by optimizing makespan and mean flow-time simultaneously. Note that, for both cases we have used the same results achieved by our approach.

The values provided in Table 3 show the makespan of the best schedules obtained in case of mt06, mt10 and mt20 problems by some GA-based scheduling algorithms. The column labeled sGA is based on the GA using the simple mutation that swaps the positions of two random jobs (Ombuki and Ventresca 2004), whereas SGA is based on simple GA proposed by Yamada and Nakano (1992, 1997). LSGA and GTGA indicate the GA-based job-shop scheduling algorithm incorporating local search and the GA based on GT crossover (Yamada and Nakano 1992), respectively. From this Table, it is ample clear that the proposed method is capable of producing near-optimal schedules in most of the cases. For mt06 and mt10 problems, it achieved the lower bound. In the case of mt20, JGGA and the proposed IPPX method cannot obtain the lower bound, however it outperforms the other contestant algorithms. Although some other heuristic scheduling algorithms may achieve the same result, it should be noted that the main objective of the proposed method is the use of new crossover technique for multi-objective JSSP. The comparisons with single objective JSSP is only to justify the capability of the proposed method in the case of minimizing makespan.

Table 3 Comparison of different operators in single objective context

4.4.2 Multi-objective context

Multi-objective optimization differs from single objective optimization in many ways. For two or more objectives, each objective corresponds to a different optimal solution; however none of the trade-off solutions is optimal with respect to all objectives. Table 4 shows the comparison of our proposed crossover-based method with some existing crossover-based methods in multi-objective context. The results shown in the Table indicate that, the IPPX based approach clearly outperforms the other crossover-based methods in multi-objective context in terms of both makespan and mean flow-time. From the results, it is also interesting to note that the performance of IPPX-based approach is superior to other in achieving better makespan values (which is considered the primary objective, if we consider single objective optimization) than the competing methods.

Table 4 Comparison with other operators in multi-objective context

Figure 9 graphically displays the convergence behavior of solutions obtained for la24 problem using different crossover operators PPX, GPMX, GOX and our proposed IPPX, averaging 30 different starting random seeds. In each case, all methods start with identical randomly picked initial solutions. Also it is able to populate the Pareto front faster than the existing approaches. As compared to other crossover-based methods, the IPPX-based method produced more non-dominated solutions with generations.

Fig. 9
figure 9

Comparison of different operators in finding Pareto optimal solutions for la24 problem

Figure 10 demonstrates the optimization behavior of the methods using the crossover operators over generations for la25 problem. The Fig. 10a presents the comparative observation on the optimization of the first objective make-span over generation. Though all PPX, GPMX, GOX and IPPX can optimize over generations successfully, IPPX converges faster than others after generation 90. Figure 10b presents the comparative observation for the second objective mean flow-time over generation. Overall PPX, GPMX, GOX and IPPX have steady convergence over the generations but IPPX shows much better convergence than others from the very first generation. In addition, the proposed approach clearly optimizes both of the objectives over the generations. From the figures, it can be found that from generation 1 to generation 150, the proposed method is able to optimize both of the makespan and mean flow-time.

Fig. 10
figure 10

a Makespan and b mean-flow time over generations for la25 problem

The aim of multi-objective optimization problems is to find all possible trade-offs among multiple objectives. To illustrate the convergence and diversity of the solutions, a comparative Pareto optimal solution to la25 problem obtained by PPX, GPMX, GOX and IPPX is shown in Fig. 11. From the figure, it can be observed that the IPPX-based approach can find a set of well diverse solutions along the Pareto front compared to its peers. For this reason, it is capable of finding extreme solutions and produce more suitable non-dominated solutions for the decision makers to choose the appropriate one based on different situations. In general, the proposed IPPX-based multi-objective JSSP algorithm is able to achieve consistent near-optimal solutions which are both well spread and converged.

Fig. 11
figure 11

Pareto optimal front obtained by IPPX, PPX, GOX and GPMX operators

It is clear that there are two distinct goals in MOEA: converge as close as possible to the Pareto-optimal front, and maintain as diverse a set of solutions over the Pareto front as possible. An MOEA will be considered good only if both of these two goals are satisfied simultaneously. Hypervolume (Zitzler et al. 2000; Veldhuizen and Lamont 2000) provides a qualitative measure of convergence as well as diversity. This metric calculates the volume in the objective space covered by members of the Pareto-optimal set with a reference point. The reference point can simply be found by constructing a vector of worst objective function values. Obviously, maximization of the volume means the minimization of all the objectives. Figure 12 shows the hypervolume enclosed by the non-dominated solutions by the operators GPMX, GOX, PPX and IPPX for la25 problem. The chosen reference point was w * = (1003.5, 860).

Fig. 12
figure 12

Hypervolume enclosed by the non-dominated solutions by different operators

The values of hypervolumes for this problem are 1068.5 for GPMX, 847.5 for GOX, 717.5 for PPX and 1784.5 for IPPX operators. Zitzler (1999) defined a metric, called maximum spread, measuring the length of the diagonal of a hyperbox formed by the extreme function values observed in the non-dominated set. The maximum spreads of the Pareto-front solutions of the corresponding operators were 40.4969 for GPMX, 48.1041 for GOX, 39.6232 for PPX, and 48.7647 for IPPX. Though the maximum spread does not reveal true distribution of the solutions, but inspecting the Fig. 12 we can see all non-dominated solutions are nearly uniformly spaced. It is evident from the above performance metrics that IPPX outperforms its peers.

4.5 Computational time

As we remove redundancy at the tail of a chromosome while applying IPPX, it finds near optimal solutions in less time duration than that of IPPX. Table 5 depicts this scenario. If we observe this table closely, we can find that the time difference increases with problem domain. In la27 (20 × 10) time difference between PPX and our approach is 4 s. It can also be noted that the time difference increases with generations in the same problem domain. In short, the larger the generation and the problem domain, the more the time difference among the proposed and the existing methods.

Table 5 Computation time by different operators

5 Conclusion

This paper presents an improved PPX-based crossover for multi-objective JSSP that reduces the redundancy at the tail of a chromosome. Experimental results exhibit that removing redundancy at the tail of a chromosome is helpful for getting better result in a reasonable computational time. Considering the real world demand for scheduling with reasonable time limit, experimental results justify that the proposed method will be very effective for large problems like 100 × 200, 150 × 300, etc. which is very usual in practice. We also believe that removing redundancy at the tail of a chromosome will be beneficial for all operation based chromosome representation for multiple-objective JSSP.