1 Introduction

Performance of assembly, production, and logistic systems are strongly dependent on the decisions made at their early design stages [1,2,3]. These systems operate based on dynamically interacting processes using soft/physical resources. Such systems are considered complex due to having large-scale and multidisciplinary design knowledge (design knowledge refers to the available information after requirement analysis) [4]. In this paper, the described systems are called ‘complex engineering process-based system’ (CEPS).

Some problems in the design of CEPSs are stated as aimed to address them by the introduced approach in this paper.

  1. 1-

    Inconsistency between different design disciplines

  2. 2-

    Incomprehensive design assessment/validation due to not having a holistic-integrated design approach.

    1. 2-1

      Assessment without considering the impact of dynamic interactions on a CEPS’s performance.

    2. 2-2

      Assessment against eighter qualitative or quantitative criteria and not both concurrently

  3. 3-

    Domain specific (not generic) approaches

Four overall phases can be defined in the design of CEPSs, requirements definition, conceptual design, configuration design, and detail design. It is not very easy to draw a line between design phases. The early-stage design, which is the scope of this paper, refers to the conceptual design stage.

The complex structure of CEPSs usually makes the designers to design different aspects of a CEPS sequentially [5]. As a result, the design at a higher-level is mostly finalized before proceeding to the next level [6]. In this context, operational policies are mostly designed after the physical design is finalized. For example, in designing a warehouse, designers decide on narrow aisles to reduce the space cost. This does not allow storing the new and picking the existing items in the same aisle concurrently (requires wide aisles). Concurrent storing-picking is an operational policy that is called double-command [7]. The operation time in the double-command process is mostly shorter than single-command, so fewer operators may deliver the same throughput. Therefore, a wide-aisle warehouse with the double-command-mode can lower the total cost. Generally, sequential approaches (not integrated) do not completely reflect the impact of single decisions on overall performance due to not having a holistic approach in addressing all types of design requirements simultaneously.

The performance of a CEPS is a function of the dynamic interactions between its resources, which can hardly be formulated with analytical methods (e.g., optimization). Therefore, it is necessary to use methods, such as simulation, that can capture the dynamic aspect of a CEPS to assess the impact of design decisions on its performance. However, simulation is often used at later design stages when important decisions are already made [8]. Integrated design in this paper means assessing the impact of design decisions on a CEPS’s performance as a function of the dynamic interactions between its resources.

Design of a CEPS is mostly subjected to multiple objectives satisfaction; quantitative (e.g., cost) and qualitative (e.g., safety). Optimization is the most applied approach for design assessment against quantitative objectives. However, the latter does not fit well with optimization methods [9]. The different nature of conflicting objectives further complicates the assessment/selection of better design alternatives. These problems are magnified due to the gap between academic approaches and what practitioners exercise [10]. Companies may develop specific approaches, which are not generic enough to be applicable in other domains.

A set of objectives are stated and their satisfaction by application of the introduced approach manifests the contributions of this work to this research context.

  • I. Enabling integrated design by.

    1. a.

      holistically addressing all types of design requirements

    2. b.

      observing the design decisions’ impact on the system’s dynamic behaviour.

  • II. Enabling multidisciplinary design to achieve design consistency.

  • III. Enabling multi-aspect design assessment by supporting the decision-making regarding the selection of better alternatives by considering different types of performance criteria.

  • IV. Having a generic design approach such that the framework is applicable in different domains.

2 Literature review

Designers devise models as their artefacts to analyze a specific design knowledge and set up a solution [11]. In CEPS design, modelling can serve three purposes; developing system architecture, design assessment, and validation [12]. System architecture embodies the system structure and function. Validation establishes evidence to assure an alternative accomplishes objectives [13]. This section reviews those research works that developed a modelling approach for CEPS design and did not focus on one modelling aspect of CEPSs, as shown in Table 1.

Table 1 Classification of modelling approaches

2.1 System architecting and validation

Different studies [14, 15, 17, 24, 26] used various methods for system architecting such as Object-Oriented (OO) or Function-Modelling (FM), but provided limited architecting guidelines.

The following works [14, 15, 24] used simulation-based validation while developing the architecture with languages, such as SysML/UML, which have simulation limitations [32]. Object-Process-Methodology (OPM) formalism offers simulation capabilities but has shortcomings in capturing system’s dynamic state at a specific time. Although, it is tried to bridge this gap by Model-to-Model transformation (MtM), working with transformation engines requires high-level programming skills [17]. Thiers [7] introduced a design methodology to support the analysis of logistics systems by using model-based systems engineering. An abstract model of a token flow network was developed to realize the MtM transformation between SysML and Petri Net (PN). However, the full automation of MtM transformation was not achieved. Moreover, no explicit approach for system architecting was suggested. The methodology could not address different control and planning policies in one analysis model and suggested modelling the system behaviour in separated behavioural diagrams.

Koo developed a modelling language for system architecting, named Object-Process Network (OPN) [27]. OPN was based on the core concepts of PN and intended to support system architecting by automating communication and computational tasks in architectural reasoning. However, OPN had shortcomings in providing explicit mechanisms for automatically generating alternatives and modelling the constraints that were related to the integration of a system’s entities. Moreover, OPN was not an intuitive language and had shortcomings in representing the static relationships between a system’s entities.

Meng [14,15,16] introduced an approach to model a Reconfigurable Manufacturing System (RMS) by application of coloured timed object-oriented Petri Nets. The approach used the OO method to identify the elements in an RMS and suggested four classes as basic elements: material processing equipment, storage equipment, robot, and Automatic Guided Vehicle (AGV). From the OO and domain modelling perspective, the defined classes were not comprehensive enough to illustrate a generic RMS. Generally, all RMSs do not use AGV and robots. Moreover, operational procedures were not addressed in that modelling approach. Also, it was not properly indicated how to formulate the problem to achieve a rigorous system configuration, such as equipment number. The presented work basically used OO at the solution level after the system architecture was finalized.

Silva and Alves [33] introduced a methodology for manufacturing systems design that produce a family of similar products. The methodology consisted of three phases: generic design, conceptual design, and detail design. This methodology was descriptive and did not suggest application of any method for system architecting, formulating solutions, and analysing the results.

Existing works mostly modelled operational policies as a fixed part of models, so simulation did not clarify the impact of different policies on performance. Abdoli and Kara [31] introduced architecting guidelines by incorporating the OO and Finite-State-Machine (FSM) methods to develop an executable system architecture (explained shortly).

2.2 Design assessment

Some works addressed quantitative and some qualitative assessment. Some studies [28, 29] used axiomatic design (AD) theory for architecting and assessment. However, AD gives less attention to design validation [34]. Cochran et al. [28, 29] introduced a methodology for designing manufacturing systems by application of AD, called: Manufacturing System Design Decomposition (MSDD). Return on investment was considered as the primary Functional Requirement (FR-1). Design Parameter-1(DP-1) was defined as ‘Manufacturing system design’ to fulfil FR-1. The next levels’ FRs were defined as measures of performance to maximize revenue. DPs were associated with FRs at each level. MSDD mainly provided the means to assess the design alternatives and did not address the design validation.

Dauby and Dagli [18] proposed an assessment approach by defining ‘canonical design primitives’ as possible genres for physical components and fuzzy logic was used to assess the impact of switching between primitives. That work only addressed the quantitative and not qualitative performance criteria. Moreover, the impact of interactions between design primitives on the system’s performance was not addressed. Kulak [30] used fuzzy logic for equipment selection, but the impact of equipment interactions with other system elements was not considered. Sasaki and Gen [35] used fuzzy logic in optimization formulation, in which the objectives and constraints were modelled as fuzzy variables. The research used GA as the search algorithm and encoded a chromosome that represented the generalized upper bounding for fuzzy objectives and variables. That approach only addressed the quantitative objectives. Moreover, that approach did not address the associated uncertainty with design choices.

Singh and Dagli [19] and Pape et al. [21] introduced similar approaches for evaluating system alternatives, called System of Systems (SoSs). The individual systems could decide whether to participate in an SoS or to make an interface with other individuals. Therefore, different SoS alternatives varied in the participation of individual systems or in developed interfaces. The performance characteristics of an SoS were defined by qualitative criteria, such as affordability. In the CEPS design context, it is needed to allocate a design option to each design requirement (such as machining equipment). Therefore, the decisions of participation or making interfaces were not equivalent to developing different alternatives in the CEPS design context. Other similar works assumed that individual systems might have priorities that affect their choice of participation [22, 23]. Accordingly, individual systems could negotiate with an SoS manager regarding deadlines and funding. The individual systems and SoS manager were modelled as agents and a fuzzy decision analysis was developed to perform the negotiations. Although, this approach resulted in a more comprehensive assessment, still had similar shortcomings to previously described work. Singh et al. [20] used fuzzy-Analytical Hierarchy Process (AHP) for decision-making on selection of manufacturing system types among; reconfigurable, dedicated, and flexible. Fuzzy logic was used to model the ratio scales because AHP has shortcomings in capturing the tied uncertainty into subjective judgments [36].

Some studies developed decision support frameworks for the selection of the most environmentally friendly supplier among existing ones by considering their logistic activities and processes performed to select the supplier that leads to a less environmental impact [36, 37]. These works compared different strategic options and did not address the system design in terms of configuring different options to develop and assess system alternatives.

2.3 Research gap

The following shortcomings are identified in reviewed works [12]. Figure 1 demonstrates how the research gaps led to a lack of practical design frameworks and consequently led to explained problems in design practices, shown on the left-hand side of Fig. 1. The right-hand side of Fig. 1 shows how the realization of objectives contributes to addressing the practical design problems because this paper aims to present a framework that is practical and can be used in real design applications.

  1. i.

    Demonstrating the system architecture and embodying dynamic behaviour in separated modelling formalisms

  2. ii.

    Modelling operational policies as model invariants

  3. iii.

    Lack of an integrated generic modelling framework that: 1) addresses different aspects of modelling at early design stages, 2) provides a systematic interconnection between different models to interchange information.

Fig. 1
figure 1

Relation between problems, research objectives, and gaps

Besides a recent work by Abdoli and Kara [31], the reviewed works realize some modelling aspects of CEPS design but not all in an integrated manner [18]. However, integrating different models enhances the consistency of the exchanged data/information [38]. Algorithmic approaches have not been used efficiently, while they can help in performing repetitive tasks [24]. The design process is knowledge propagation from one model to another. Thus, an integrated modelling approach can overcome the design methodologies’ shortcomings that mainly guide the design process without generating solutions. Different modelling approaches answer narrowly defined design questions. Sequential design approaches also hardly integrate various design stages or capture the connection between low-level design decisions and high-level system objectives [39]. This highlights the importance of having an integrated design approach [40].

Model-based design is a way of formalization by defining the semantics of representing something before describing it. Model-based design relaxes the application of framework from the CEPS domain, assuring it is domain neutrality [16].

There is a lack of a generic modelling approach providing prescriptive guidelines for system architecting, multi-aspect assessment, and validation, which can assist in the integrated design of CEPSs [16, 41].

Modelling is both art and science, which conceptual modelling lying more at the artistic end [42]. Two designers may use one modelling approach and resultant models can be different with different capabilities. A prescriptive approach reduces the modelling effort and supports the design process more effectively [43]. The proposed framework utilizes different methods and supporting algorithms and systematically integrates them. The framework prescribes how to develop models (by introducing novel approaches for methods’ application) and algorithms. Each model/algorithm serves a specific role in this framework. Systematic interconnection between models and algorithms allows a smooth inter-operability between them to serve framework objectives. Contributions of the proposed framework rely on how the methods are employed and the systematic models’ integration. Each of the applied methods/formalisms may (may not) have been used before for other narrowly defined design purposes, yet the prescribed approaches in this framework are novel and contribute to satisfying research objectives.

3 Framework structure and applied methods

An overview of the framework and justification of methods selection is provided in this section. This research used Systems Engineering (SE) for defining the framework structure, in terms of its tasks and their logical sequence, due to the multidisciplinary nature of CEPSs. Because SE encompasses interdisciplinary activities and concentrates on system properties rather than on single disciplines [44]. In SE process, after requirement analysis, the system architecture is developed [44], which is equivalent to design at early stages. In the trade study step, design alternatives are configured, their performance is studied, and one is nominated for detailed design. V-model of SE gives high attention to design validation from early stages, conforming to the importance of observing the dynamic behaviour at early stages by interconnecting the integration-validation phase with architecture design. Hence, the proposed framework addresses: architecture design, validation, and trade study, see Fig. 2.

Fig. 2
figure 2

The overall structure of the framework with applied methods and their data exchange flow

The framework develops an executable architecture for observing the behaviour (dynamic performance) of alternatives by simulation. However, design is a process of design generation, evaluation, and redesign [45]. The proposed trade study mechanism serves a multitude of alternative generation, design constraints realization, and assessment. The OO modelling is a well-known approach for designing complex systems with multi- hierarchal structures [46]. Hence, the framework uses the OO method and architecting guidelines are developed for capturing the complexity of holistically modelling CEPSs.

In the framework, the validation happens at the system level. Discrete-Event-Simulation (DES) is a promising method for validation of the dynamic aspects of CEPSs [40]. PN is a widely used formalism in this context. However, a PN-model structure can look complicated even for small-scale systems [14, 47]. The FSM represents system structure by dynamic states of system elements. Hence, the framework maps the mentioned ‘architecting guidelines’ into FSM formalism to develop an executable architecture.

Design knowledge database in this framework stores design options with their properties, called design-objects. ‘Alternative generation’ algorithm configures the design alternatives by allocation of design-objects to design requirements while the ‘Feasibility checking’ algorithm crosses off the infeasible alternatives.

The framework uses optimization/quantification models to develop a numerical model for alternatives, which are later simulated for validation. The framework uses Linear-Programming (LP) and Automated Layout Design Program (ALDeP). The latter is only used if layout design is needed. Other optimization methods can also be used because the analytical formulations of CEPSs belong to NP-hard problems and no optimization method can guarantee to find the global optimum [48]. The proposed framework used LP because of its broad application field including scheduling, and resource allocation. For layout development, CORE or CRAFT can also be used instead of ALDeP.

Many studies argue that qualitative properties are more critical for a system’s lifetime value [49]. Yet, qualitative aspects (e.g., safety) are not very easy to be assessed by optimization models due to the inherited uncertainty in qualitative assessment [50]. In the literature, combinations of qualitative assessment methods are used such as fuzzy-AHP or fuzzy-Quality Function Deployment to assess the qualitative aspects of a CEPS [51]. However, fuzzy logic played the key role in capturing the uncertainty in qualitative assessment. The framework includes a fuzzy assessment approach to assess the alternatives against qualitative objectives.

This framework uses simulation results as one and fuzzy assessment as another source of information in the trade study step. A feedback algorithm is introduced to modify an alternative according to the simulation results. TOPSIS (Technique for Order of Preference by Similarity to Ideal Solution) is widely used in Multi Criteria Decision Making problems because it allows different criteria have their own measuring units. TOPSIS also considers different importance for criteria. This framework uses TOPSIS to rank the alternatives and identify the final alternative that demonstrates better performance by considering simulation and fuzzy assessment results. The framework is implemented in MATLAB, however other platforms can be used too.

Examples of a warehouse design are given in the following sections to explain the methods’ applications and the connection between different artefacts in the framework. However, this does not impact the domain neutrality of the framework, as its application is presented in two case studies from different domains later in the paper.

4 Developing an executable system architecture

Abdoli and Kara [31] introduced architecting guidelines using the OO method for holistically modelling the architecture of a CEPS (decomposition hierarchy). Accordingly, a CEPS is decomposed into processes that deliver the main function of a CEPS, as the transformation of an item’s state from input to output. For instance, warehousing processes can be defined as; receiving, storing, picking, and shipping [52]. Each process can have any type of physical/operational enablers with their Design Requirements (DRs). The architecture is constructed in FSM formalism to achieve an executable architecture and is named System-State Meta-Model (SSMM). These architecting guidelines are briefly explained here and interested readers are referred to this reference [31] for details.

The OO-based hierarchal decomposition approach is mapped into FSM formalism by embodying nested states in SSMM, an example is shown in Fig. 3. In this approach, a CEPS is firstly modelled as a parent state in FSM and then is decomposed into sub-states to embody the CEPS’s processes, called process-state. Each process-state is decomposed into further sub-states that represent its needed enablers, called enabler-state. Further, each enabler-state is decomposed into sub-states, which address its specific DRs, called DR-state. The possible design options/genres for each DR-state are modelled as its sub-states, called object-state. The developed SSMM in FSM realizes a holistic design approach by embodying all types of DRs. For the warehouse case, enablers of storing process are defined as operational policy, equipment, and infrastructure. For instance, it is a DR to define an operational policy for allocating the Stock Keeping Units (SKUs) to storage modules. Hence, ‘SKU-Allocation’ is defined as a DR-state, which includes two sub-states showing two possible operational policy options: ‘Class-Based Storage’ (CBS) and ‘Random’.

Fig. 3
figure 3

SSMM for warehouse-case study

FSM allows defining functions for a state to model its behaviour. In this framework, a function is associated with each object-state to model its dynamic behaviour, called object-functions. An object-function returns the dynamic state of an object-state with certain outputs (e.g., equipment available capacity), called Dynamic Variables (DVs). Relation between object-functions (in exchanging DVs) must be defined to embody the interactions between enablers.

In SSMM, a design alternative is configured by activating one object-state for a DR-state. For instance, one alternative is configured by activating ‘CBS’ policy state while another is configured by activating ‘Random’ state. This model is named SSMM because it demonstrates system architecture by states while it is a meta-model that various system alternatives conform to it.

5 Trade study process

5.1 Alternative generation and feasibility checking algorithms

The alternative generation algorithm generates alternatives by exploring the stored design-objects in the design knowledge database [53]. The algorithm uses an array of variables, which each represents a DR with a unique ID, see Eq. 1. In this equation, Z and K are respectively the numbers of defined DRs and design-objects. The feasibility checking algorithm eliminates those alternatives that combination of employed design-object results in an unfeasible system due to inconsistencies between their disciplines.

The applicability attribute of a design-object clarifies whether it can be employed for a \({\text{DR}}_{{{\text{ID}}}}\). The consistency attribute clarifies the consistency/compatibility of a design-object with others. Each design-object has an identifier, called Uniqueness-Key (UK). Generated alternatives at this stage are called ‘qualitative’ because they only show the employed genres for DRs.

$$ {\text{CEPS}}_{{{\text{Qualitative}}\;{\text{reference-}}{\text{model }}}} = \left[ {\left( - \right)_{{{\text{DR}}_{1} }} ,.,\left( - \right)_{{{\text{DR}}_{{id}} }} ,.,\left( - \right)_{{{\text{DR}}_{z} }} } \right] $$
(1)
$$ {\text{Applicability}}_{{{\text{UK}}}} { } = ({\text{Applicability}}_{1} ,{ } \ldots ,{\text{Applicability}}_{{\text{z}}} ){ } $$
(2)
$$ {\text{Applicability}}_{{{\text{UK}},{\text{id}}}} = \left\{ {\begin{array}{*{20}c} {1;Can be used in {\text{DR}}_{{{\text{id}}}} } \\ {0;Otherwise} \\ \end{array} } \right.\forall {\text{id}} \in {\text{Z}} $$
$$ {\text{Consistency}}_{{{\text{UK}}}} = \left( {{\text{Consistency}}_{1} , \ldots ,{\text{Consistency}}_{{\text{k}}} } \right) $$
(3)
$$ {\text{Consistency}}_{{{\text{UK}},{\text{ j}}}} = \left\{ {\begin{array}{*{20}c} {1;Is consistent with {\text{design}} - {\text{object}}_{{\text{i}}} } \\ {0;otherwise} \\ \end{array} { }\forall {\text{i}} \in {\text{k}}} \right. $$

5.2 Optimization/quantification

Optimization models are applied to qualitative alternatives to achieve a level of quantification, which with can be simulated. As explained, this framework uses LP and ALDeP for quantification of alternatives [31]. Quantification models return the required numbers of an employed design-object to satisfy objectives/constraints (e.g., three forklifts are required to meet throughput), as shown by ‘Number of Instances’ in Eq. 4. Figure 4 demonstrates how this framework characterizes the solution space. The quantified alternatives are stored in the design database.

$$ \begin{aligned}& {\text{CEPS}}_{{{\text{Quantitative}} - {\text{Alternative}}_{{\text{t}}} { }}}\\ &\quad = \left[ {\left( { - ,{ } - } \right)_{{{\text{DR}}_{1} }} ,., \left( {{\text{UK}},{\text{ Number of Instances}}} \right)_{{{\text{DR}}_{ID} }} ,., \left( { - ,{ } - } \right)_{{{\text{DR}}_{z} }} } \right] \end{aligned}$$
(4)
Fig. 4
figure 4

Characterizing the solution space in the process of searching for a better design alternative

5.3 Fuzzy assessment framework

This framework utilizes a novel fuzzy framework for assessment of alternatives against qualitative objectives by modelling the domain expert’s opinion [54]. The fuzzy framework has a hybrid-hierarchal structure and comprises fuzzy models in three levels: design-object, process, and system. The fuzzy models with respect to their hierarchal order are called ‘object-fuzzy-model’, ‘Process-fuzzy-model’, and ‘System-fuzzy-model’, see Fig. 5. The fuzzy framework has a hierarchal structure such that the output of a lower-level fuzzy-model is an input to the next-level model. The fuzzy framework has a hybrid structure, so the fuzzy models get extra inputs related to each assessment level. Hence, a fuzzy model at a higher level has at least two input sets: output of a lower-level fuzzy model and extra inputs from the design knowledge.

Fig. 5
figure 5

Fuzzy Assessment framework structure

5.3.1 Fuzzy framework structure

The rule-based fuzzy structure is used in the fuzzy framework because it fits well to decision-making in the presence of linguistic information. Also, the Mamdani approach is used as the inference system because it fits better into rule-based models and returns a single output [55]. The novelty of the introduced fuzzy framework relies on its structure (hybrid-hierarchal), hence the technical details of fuzzy logic are not explained and interested readers are referred to literature [55].

In rule-based models, a set of if–then rules maps inputs to outputs [56]. Because the fuzzy framework aims to assess the qualitative aspects of a CEPS, so linguistic labels are defined for fuzzy inputs/outputs. The framework uses 'Efficiency' labels to demonstrate the input fuzzy sets and 'Goodness' labels for outputs, which Input labels are labelled as ‘Not-Efficient’, ‘Marginally-Efficient’, ‘Efficient’, and ‘Very-Efficient’, and output labels as: ‘Not-Good’, ‘Marginally-Good’, ‘Good’, and ‘Very-Good’.

In the presence of linguistic information, experts define if–then rules and grade membership functions such that the model returns satisfactory results with a test dataset [56]. For instance, if the efficiency of a design-object against a ‘very-important’ objective is ‘Very Efficient’ and against a ‘less-important’ objective is ‘Marginally Efficient’, the goodness of design-object is assessed: ‘Good’.

This paper uses the triangular function, which is extensively used in the application of fuzzy logic in different contexts. The framework uses the following widely used procedure introduced by this reference [55] to grade the membership functions. It includes three steps that should be followed for each fuzzy variable:

  1. (1).

    Defining a range that indicates the discourse universe of a fuzzy variable; for example (0–1).

  2. (2).

    Experts mark an interval in favour of the meaning of each linguistic label used for the variable. This interval is defined (0.45, 0.8) for ‘Efficient’ label, see Fig. 5.

  3. (3)

    Experts mark a key-point within the defined interval, significantly representing the meaning of the linguistic label. The membership function gets its maximum truth degree in key-point. This is defined (0.75) for ‘Efficient’ label.

The membership function associated with each fuzzy set covers a sub-range in the discourse universe. For example, ‘Not-Efficient’ function covers (0–0.35) subrange while ‘Marginally-Efficient’ covers (0.25–0.6). The same fuzzification is applied for variables that are exchanged between fuzzy models. This novel structure facilitates the data exchange between fuzzy models because both input and output variables embody the same fuzzy magnitude about the domain’s expert opinion. For instance, fuzzification of ‘Marginally-Efficient’ is similar to ‘Marginally-Good’.

All fuzzy models take the objectives’ importance as a fuzzy input too. The fuzzy framework is implemented using MATLAB fuzzy toolbox as an “off-the-shelf” platform, used in different works [57]. However, other platforms can be used too.

5.3.2 Fuzzy models

‘Object-fuzzy-model’: A CEPS performance is directly related to the employed enablers for DRs. The ‘Object-fuzzy-model’ assesses the goodness of a design-object to be employed in an alternative. Thus, the domain expert’s opinion about the efficiency of a design-object against objectives is modeled as a fuzzy input for ‘Object-fuzzy-model’.

This framework uses a procedure to assist the domain experts to express their opinion regarding the efficiency of fuzzy inputs in all three fuzzy models:

  1. (1)

    The expert expresses their opinion about an input by selecting one efficiency label among: ‘Not-efficient’, ‘Marginally-Efficient’, ‘Efficient’, ‘Very-Efficient’.

  2. (2)

    The expert picks a value from the sub-range covered by the selected efficiency label. For instance, if the expert selects ‘Efficient’, then they pick a value from the subrange of (0.45–0.8).

In the warehouse case, 'Cargomatic' is a design-object that can be employed for \({\mathrm{Unloading}-\mathrm{equipment type}}_{1}\) as \({\mathrm{DR}}_{1}\). Objectives are safety satisfaction and future expansion; the former is considered as very-important and the latter as less-important. The expert assesses ‘Cargomatic’ as ‘Efficient’ against safety satisfaction and picks 0.55 from the covered sub-range by ‘Efficient’ membership function. This value (0.55) belongs to two membership functions with different truth degrees: ‘Efficient’ and ‘Marginally-Efficient’. This conforms to the main concept of fuzzy logic. Although the idea was to assess ‘Cargomatic’ as ‘Efficient’, multiple belongingness of 0.55 to ‘Efficient’ and ‘Marginally-Efficient’ embodies uncertainty in subjective assessment. It is worth mentioning that instead of using the above procedure, it is also possible to use the introduced approach by this reference [58] to calculate the assessment results in all three fuzzy models. Hence, the experts only pick an efficiency label (such as ‘Very-Efficient’). Then the introduced FIS can be used to either return a single value or a linguistic term as fuzzy assessment output.

‘Process-fuzzy-model’: A process may need to have multiple interacting enablers to perform its function. ‘Process-fuzzy-model’ assesses the goodness of a process as the configuration of the employed design-object that interact with each other. Thus, the first input set to ‘Process-fuzzy-model’ is the goodness of design-objects (output of ‘Object-fuzzy-model’). The expert’s opinion about the interaction efficiency between every two design-objects is modelled as the second fuzzy input for the ‘Process-fuzzy-model’. For example, the interaction efficiency of ‘Cargomatic’ and ‘Sunken-door’ against safety satisfaction is expressed as ‘Very-Efficient’, while interaction efficiency of ‘Cargomatic’ and ‘Level-door’ is expressed as ‘Not-Efficient’.

Experts define if–then rules. For instance, if the goodness of all employed design-objects is 'Very-Good' and interaction efficiency of any employed design-object against very-importance objective is not 'Not-Efficient' then the process is assessed as 'Very-Good'.

‘System-fuzzy-model’: At the system level, certain processes can be more important due to their major impact on objectives’ satisfaction. In the warehousing domain, the storing process is very critical and its design directly impacts the warehouse performance [53]. Thus, three input fuzzy sets are defined for ‘system-fuzzy-model’: ‘Goodness’ of processes (output of ‘Process-fuzzy-model’), processes’ importance, and objectives’ importance. ‘System-fuzzy-model’ assesses an alternative as an integration of processes. For instance, if the goodness of very-important processes is 'Not-Good' and the goodness of less-important processes is not 'Very-Good' then the ‘System-fuzzy-model’ assesses the alternative as 'Not-Good'.

The fuzzy framework provides a single value as a design goodness score, which a higher score means a better alternative. The framework retrieves its needed data from the design knowledge database. For example, the design-objects have their interactions’ efficiencies as their attributes.

5.4 System validation

The number of design alternatives and DRs can be massive. Accordingly, configuring an alternative in SSMM by activating the employed object-states in DR-states can take so much time. Hence, it is needed to facilitate such a tedious task to assure the practicality of the framework. On the other hand, the simulation results can provide some insights into the dynamic behaviour of the alternatives, which can be used in a systematic manner to improve them. To address these issues, two supporting algorithms are introduced: ‘coupling algorithm’ and ‘feedback algorithm’.

5.5 Coupling algorithm

‘Coupling algorithm’ retrieves the quantified alternatives from the database and simulates them with SSMM automatically. This is called coupling algorithm because it configures the SSMM such that represents one specific alternative. Generally, alternatives vary in their employed design-objects for DRs. For each specific alternative, the coupling algorithm activates the object-states equivalent to its employed design-objects. For ease of referring, the alternative that is going to be simulated is called ‘current-alternative’.

In SSMM, each DR-state has a constant value called ‘DRID’, equivalent to the ID of \({\mathrm{DR}}_{\mathrm{ID}}\). For instance, \({\mathrm{Storing}-\mathrm{Allocation}}_{8}\) is a DR and its ID is 8. Accordingly, ‘SKU-Allocation’ state is equivalent to \({\mathrm{Storing}-\mathrm{Allocation}}_{8}\) and the DRID of ‘SKU-Allocation’ is also valued 8.

As explained, each design-object has an identifier called Uniqueness Key (UK). The equivalent object-state to a design-object carries a constant with the same value as UK. For instance, CBS is a design-object, and its UK is valued 13. Likewise, CBS-state carries 13 as its constant value.

The coupling algorithm retrieves the quantitative model of current-alternative from the database and obtains the UK of the employed design-object for each \({\mathrm{DR}}_{\mathrm{ID}}\). For the warehouse case, the current-alternative employed CBS as its design-object, which its UK is equal to 13, as shown below.

$${\mathrm{Warehouse}}_{\mathrm{Quantitative\,model }}=\left[{\left(-\right)}_{{\mathrm{DR}}_{1}},.,{\left(13\right)}_{{\mathrm{DR}}_{8}},.,{\left(-\right)}_{{\mathrm{DR}}_{z}}\right]$$

In SSMM, a transition condition is defined for each object-state to compare the retrieved UK by coupling algorithm with the carrying constant by an object-state. In the warehouse case, the algorithm finds that the employed design-object for \({\mathrm{Storing}-\mathrm{Allocation}}_{8}\) has UK of 13. There are two object-states in the ‘SKU-Allocation’ state: ‘CBS’ and ‘Random’. The transition condition in ‘CBS’ compares the retrieved UK (13) with the carrying constant by CBS, which is also 13. Therefore, the transition happens, and CBS-state is activated. While the ‘Random’ state carries 14 as its constant. When the transition is tested in the Random state, since the retrieved value for \({\mathrm{Storing}-\mathrm{Allocation}}_{8}\) is 13, not equal to 14, the transition does not happen.

The coupling algorithm checks each DR-state and activates the employed design-objects for each DR-state. At this stage, SSMM demonstrates a specific alternative that can be simulated.

Coupling algorithm:

Find quantitative model of an alternative for simulation.

For id = 1: Z (number of DRs).

  1. 1.

    Find ‘DRID’ in \({\mathrm{DR}}_{\mathrm{id}}-\mathrm{state}\) equivalent to \({\mathrm{DR}}_{\mathrm{id}}\)

  2. 2.

    In quantitative model, retrieve UK of employed design-object for \({\mathrm{DR}}_{\mathrm{id}}\).

  3. 3.

    Check transition conditions of object-states in \({\mathrm{DR}}_{\mathrm{id}}-\mathrm{state}\)

  4. 4.

    Activate the object-state that its constant is equal to retrieved UK

End.

5.5.1 Feedback algorithm

It is critical that a design alternative satisfies objectives and constraints (e.g., cost and throughput), so they are formulated in SSMM and are called Design-KPIs. The object-functions dynamically update these variables and return their value at the end of simulation.

Due to the dynamic infarctions between the enablers (more accurately the selected design-objects for DRs), the simulation may return different values for the design-KPIs compared to the calculated results by quantification models. For instance, the simulated throughput can be lower than what was formulated as an LP constraint. Likewise, the total cost can be higher than the LP result. Moreover, an alternative may have the potential for better performance (e.g., throughput can be satisfied with lower cost). This paper introduces the ‘feedback algorithm’ to study the simulation results and modify the quantification of alternatives to assure fulfilment of constraints or improve objectives’ satisfaction.

This paper recommends formulating the ‘queue’ between the processes and ‘utilization’ of physical enablers as ‘dynamic-KPIs’ in SSMM. The utilization is formulated for those enablers whose limited capacity may affect objectives satisfaction. Because optimization approaches mostly model a system such that its enablers use their nominal capacity. However, the different operation times of different enablers may lead to a desynchronization between processes. Consequently, enablers may run below their nominal capacity, impacting the throughput. The queue between two processes is formulated by considering the finish time of items in the first process and their start time in the subsequent process. The utilization of enablers is formulated by considering the accumulated values of their busy times during the available working time.

The feedback algorithm includes two procedures. The first addresses modifying the alternatives to meet the throughput if it was not satisfied according to simulation results. The second procedure aims to reduce the number of physical enablers without compromising on throughput satisfaction to improve the alternatives from the cost reduction perspective.

If an alternative does not satisfy the throughput, the process that experiences a long queue (bottleneck) is modified by increasing the number of the physical enabler (e.g., machining equipment) that its limited available capacity caused the bottleneck situation. That is identified by studying the utilization of the physical enablers in the bottleneck process. Hence, the enabler with the highest utilization is the first candidate to increase its number of instances.

If the utilization of an enabler in one process is considerably low, the number of instances for that enabler is reduced, because the process might be able to meet throughput with a smaller number of enablers.

Feedback algorithm:

  1. A)

    Throughput constraint satisfaction procedure

    1. I.

      Select Bottle-Neck process (BN).

    2. II.

      Define ‘Difference’ as: ‘required-throughput’ – ‘simulated-throughput’

    3. III.

      In BN, find enabler(s) that had highest simulated utilization; call it \({\text{enabler}}_{{\text{x}}}\)

    4. IV.

      Find ‘Number of Instances’ for \({\text{enabler}}_{{\text{x}}}\) in quantitative model of alternative

    5. V.

      ‘Updated-Instances’ = ‘Number of Instances’ + \(\left(\frac{{{\text{Difference}}}}{{{\text{Capcityof enabler}}_{{\text{x}}} { }}} \times {\text{operation time of enabler}}_{{\text{x}}}\right)\)

    6. VI.

      Call quantification to apply needed changes according to ‘Updated-Instances’

  2. B)

    Objectives’ improvement procedure

    1. I.

      Do these steps for each enabler of a process

    2. II.

      Define: n = ‘Number of Instances’, ‘Minimum’ = ‘Number of Instances’

    3. III.

      Find Total-Busy-Time (TBT) of all instances of enabler

    4. IV.

      Perform this loop while the below equation holds

      (a) For k = 1: n-1

      $$ \begin{aligned} \frac{{\left( {\text{Avaiable time}} \right) \times \left( {{\text{n}} - {\text{k}}} \right)}}{{\left( {\text{Avaiable time}} \right) \times {\text{n}}}} & \le \frac{{\mathop \sum \nolimits_{{{\text{i}} = 1}}^{{\text{n}}} {\text{TBT of instance }}_{{\text{i}}} }}{{\left( {\text{Avaiable time}} \right) \times {\text{n}}}}\\ & \le \frac{{\left( {\text{Avaiable time}} \right) \times \left( {{\text{n}} - {\text{k}} + 1} \right)}}{{\left( {\text{Avaiable time}} \right) \times {\text{n}}}} \end{aligned} $$

      (b) If above equation holds then ‘Minimum’ = n-k + 1

    5. V.

      ‘Updated-Instances’ = ‘Minimum’

The introduced feedback algorithm addresses simple modifications to alternatives in a systematic manner to accelerate the design process and reduce mistakes. A modified alternative is simulated again. The algorithm can be called several times until the alternative satisfies the constraints or there is no place for further improving the objectives. A certain number of iterations can be defined as another stopping condition for the algorithm.

From the SE perspective, an alternative is validated at this stage. The SSMM, coupling algorithm, feedback algorithm, and quantification models work in tandem and exchange their results as shown in Fig. 2.

5.6 Technique for order of preference by similarity to ideal solution

The introduced framework provides two sources of information to analyse/assess the performance of alternatives: the simulation results for the quantitative objectives and the fuzzy assessment for the qualitative ones. This framework uses TOPSIS for ranking alternatives and final selection of a better alternative.

The TOPSIS procedure is explained briefly here and interested readers are referred to the literature for details. TOPSIS hypothesises two alternatives: positive and negative ideal alternatives. The former illustrates the best performance against all criteria while the latter demonstrates the worst. TOPSIS finds a real-alternative that is closest to the positive and is farthest from the negative ideal-alternative. The total number of feasible alternatives is defined as m while n is the total number of assessment criteria (e.g., simulated total cost and fuzzy assessment).

TOPSIS application:

  1. I

    I- Do it for i = 1: m

    1. 1.

      Let \({\text{x}}_{{{\text{ij}}}}\) be the given score of \(alternative_{i}\) against criterion \({\text{j}}\). Construct matrix \({\text{X}}\) as: \({\text{X}} = \left[ {{\text{x}}_{{{\text{ij}}}} } \right]_{{{\text{m}} \times {\text{n}}}}\)

    2. 2.

      Let \({\text{J}}\) be the set of benefit criterions, which their maximization is desired (e.g., throughput):

    3. 3.

      Let \({\text{J}}^{\prime}\) be the set of cost criterions, which their minimization is desired (e.g., cost):

    4. 4.

      Let \({\text{z}}_{{\text{m}}}\) be each criterion weight:

    5. 5.

      Construct normalized weight matrix\(: {\text{W}} = \left[ {{\text{w}}_{{\text{j}}} } \right]_{{1 \times {\text{n}}}}\), \(\mathop \sum \limits_{{\text{j}}} {\text{w}}_{{\text{j}}} = 1\); \({\text{w}}_{{\text{j}}} {\text{ = z}}_{{\text{j}}} /\mathop \sum \limits_{{\text{j}}} {\text{z}}_{{\text{j}}}\) m = 1: n

    6. 6.

      Construct normalized decision matrix \({\text{r}}\): \({\text{r}} = \left[ {{\text{x}}_{{{\text{ij}}}} } \right]_{{{\text{m}} \times {\text{n}}}} ; {\text{r}}_{{{\text{ij}}}} = \frac{{{\text{x}}_{{{\text{ij}}}} }}{{\surd \mathop \sum \nolimits_{{\text{i}}} {\text{x}}_{{{\text{ij}}}}^{2} }}\) for i = 1: m and j = 1: n

    7. 7.

      Construct weighted normalized decision matrix: \({\text{v}} = \left[ {{\text{v}}_{{{\text{ij}}}} } \right]_{{{\text{m}} \times {\text{n}}}}\), \({\text{v}}_{{{\text{ij}}}} = {\text{w}}_{{\text{j}}} \times {\text{r}}_{{{\text{ij}}}}\),

    8. 8.

      Positive Ideal-alternative:\({\text{ A}}^{*} = \{ {\text{v}}_{1}^{*} ,{ } \ldots ,{\text{v}}_{{\text{n}}}^{*} \} ;{ }\) \({\text{v}}_{{\text{j}}}^{*} = {\text{\{ max }}({\text{v}}_{{{\text{ij}}}} {\text{) if j }} \in {\text{J}};{\text{min }}({\text{v}}_{{{\text{ij}}}} ){\text{ if j }} \in {\text{J}}^{\prime}\}\)

    9. 9.

      Negative Ideal-alternative:\( {\text{A}}^{\prime } {\text{ = }}\{ {\text{v}}_{1}^{\prime } {\text{\_}}, \ldots ,{\text{v}}_{n}^{\prime } \} \),\( {\text{v}}_{j}^{\prime } = \{ {\text{min}}({\text{v}}_{{{\text{ij}}}} {\text{)}}\;{\text{if}}\;{\text{ j}} \in {\text{J}};{\text{max}}({\text{v}}_{{{\text{ij}}}} )\;{\text{if}}\;{\text{ j}} \in {\text{J}}^{\prime } \} \)

    10. 10.

      Separation from Positive Ideal-alternative is: \({\text{S}}_{{\text{j}}}^{*} = \left[ {\mathop \sum \limits_{{\text{j}}} ({\text{v}}_{{\text{j}}}^{*} - {\text{v}}_{{{\text{ij}}}} )^{2} } \right]^{{{\raise0.7ex\hbox{$1$} \!\mathord{\left/ {\vphantom {1 2}}\right.\kern-\nulldelimiterspace} \!\lower0.7ex\hbox{$2$}}}} {\text{ i}} = 1:{\text{m}}\)

    11. 11.

      Separation from Negative Ideal-alternative is:\({\text{S}}_{{\text{j}}}^{^{\prime}} = \left[ {\mathop \sum \limits_{{\text{j}}} ({\text{v}}_{{\text{j}}}^{^{\prime}} - {\text{v}}_{{{\text{ij}}}} )^{2} } \right]^{{{\raise0.7ex\hbox{$1$} \!\mathord{\left/ {\vphantom {1 2}}\right.\kern-\nulldelimiterspace} \!\lower0.7ex\hbox{$2$}}}} {\text{ i}} = 1:{\text{m}}\)

    12. 12

      Calculate relative closeness to ideal solution,\({\text{ C}}_{{\text{i}}}^{*}\) as: \({\text{C}}_{{\text{i}}}^{*} = {\text{S}}_{{\text{j}}}^{^{\prime}} /\left( {{\text{S}}_{{\text{j}}}^{^{\prime}} + {\text{S}}_{{\text{j}}}^{*} } \right)\) i = 1: m \(0 < {\text{C}}_{{\text{i}}}^{*} < 1\)

  2. II.

    II- Select the alternative with \({{C}_{i}}^{*}\) closest to 1.

TOPSIS takes simulation and fuzzy assessment results of alternatives and ranks them from the best to worst, which the best one is handed-off for detailed design. Hence, this framework can stand as a decision support framework at early design stages of CEPSs.

In the next section, two case studies are presented to demonstrate the applicability of the proposed framework and its advantages compared to existing approaches in the literature.

6 Case studies

6.1 Warehouse design case

A pharmaceutical company wants to construct a warehouse that will receive three SKU types: SKU-A, SKU-B, and SKU-C. Stakeholders defined the design objectives as cost minimization, safety satisfaction, and future expansion while meeting the required throughput. The warehouse managers defined the process enablers as equipment, infrastructure, and operational policy. The warehouse managers defined DRs for process enablers and their possible design-objects. This is all part of the design knowledge used as input for the framework. Warehouse SSMM is demonstrated in Fig. 3.

Generally, the warehousing function starts by receiving consignment and continues by storing SKUs. The assignment policy determines how to allocate SKUs into the storage modules. The CBS policy allocates the high-demand SKUs in modules that require less travel-time (operation time) in storing/picking processes, while this increases the travel-time for low-demand SKUs. The random storage policy allocates SKUs to any available spot. The batch and single picking are two operational policies for the order-picking process, the former consolidates orders and make a picklist. Whereas in the single-picking policy, the orders are picked separately. The travelled distance of the operator directly affects the operation times in storing/picking processes. Having a cross-aisle between storage aisles can reduce the travel-distance, because the operator can exit/enter aisles without passing the entire aisle length, yet a warehouse with a cross-aisle requires more space.

6.1.1 Alternative generation

The qualitative reference model of the warehouse is shown below, which includes 20 defined DRs.

\({\mathrm{Warehouse}}_{\mathrm{Qualitative}-\mathrm{reference Model}}\): \({[\mathrm{Unloading}-\mathrm{equipment type}}_{1}\), \({\mathrm{Unloading}-\mathrm{door position}}_{2}\), \({\mathrm{Unloading}-\mathrm{ door type}}_{3}\), \({\mathrm{Staking}-\mathrm{timing}}_{5}\), \({\mathrm{Stacking}-\mathrm{infrastructure type}}_{6}\), \({\mathrm{Stacking}-\mathrm{equipment type}}_{7}\), \({\mathrm{Storing}-\mathrm{Allocation}}_{8}\), \({\mathrm{Storing}-\mathrm{timing}}_{9}\), \({\mathrm{Storing }-\mathrm{parallel}-\mathrm{mode}}_{10}\), \({\mathrm{Storing}-\mathrm{Equipment}-\mathrm{type}}_{11}\), \({\text{ Storing}} - {\text{storage}} - {\text{module}} - {\text{type}}_{12}\), \({\text{ Storing}} - {\text{aisles configuration}}_{13}\), \({\text{ Picking}} - {\text{pick}} - {\text{listing}}_{14}\), \({\text{Picking}} - {\text{parallel mode}}_{15}\), \({\text{Picking}} - {\text{Equipment}} - {\text{type}}_{16}\), \({\text{Shipping}} - {\text{door}} - {\text{type}}_{17}\), \( {\text{Shipping}} - {\text{door}} - {\text{position}}_{18}\), \({\text{Shipping}} - {\text{timing }}_{19}\), \({\text{Shipping}} - {\text{Equipment}} - {\text{type }}_{20} ]\).

Figure 6 shows the stored design-objects in the design database with their attributes (as input data such as applicability, consistency, and efficiency) to be used by feasibility checking/alternative generation algorithms, quantification models, and fuzzy framework.

Fig. 6
figure 6

Defined design-objects for the warehouse case study

The feasibility checking algorithm crossed-off almost half of the generated alternatives, showing the high risk of design failure without having an integrated design approach. Three generated alternatives are shown for demonstration.

\({\text{Alternative}}_{{\text{A}}}\) = \([\left( 1 \right)_{1}\),\(\left( 3 \right)_{2}\),\(\left( 5 \right)_{3}\),\(\left( 7 \right)_{4}\),\(\left( 7 \right)_{5}\),\(\left( {10} \right)_{6}\),\(\left( {11} \right)_{7}\), \(\left( {14} \right)_{8}\),\(\left( 7 \right)_{9}\),\(\left( {15} \right)_{10}\),\(\left( {11} \right)_{11}\),\(\left( {17} \right)_{12}\),\(\left( {20} \right)_{13}\),\(\left( {21} \right)_{14}\),\(\left( {15} \right)_{15}\),\(\left( {12} \right)_{16}\), \(\left( 6 \right)_{17}\),\(\left( 3 \right)_{18}\),\(\left( 8 \right)_{19}\),\(\left( 2 \right)_{20} ]\).

\({\text{Alternative}}_{{\text{B}}}\) = \([\left( 1 \right)_{1}\),\(\left( 3 \right)_{2}\),\(\left( 6 \right)_{3}\),\(\left( 7 \right)_{4}\),\(\left( 7 \right)_{5}\),\(\left( 9 \right)_{6}\),\(\left( {11} \right)_{7}\), \(\left( {13} \right)_{8}\),\(\left( 7 \right)_{9}\),\(\left( {15} \right)_{10}\),\(\left( {11} \right)_{11}\),\(\left( {17} \right)_{12}\),\(\left( {20} \right)_{13}\),\(\left( {22} \right)_{14}\),\(\left( {15} \right)_{15}\),\(\left( {11} \right)_{16}\), \(\left( 6 \right)_{17}\),\(\left( 3 \right)_{18}\),\(\left( 7 \right)_{19}\),\(\left( {11} \right)_{20} ]\).

\({\text{Alternative}}_{{\text{c}}}\) = \([\left( 1 \right)_{1}\),\(\left( 3 \right)_{2}\),\(\left( 5 \right)_{3}\),\(\left( 7 \right)_{4}\),\(\left( 7 \right)_{5}\),\(\left( {10} \right)_{6}\),\(\left( {11} \right)_{7}\), \(\left( {13} \right)_{8}\),\(\left( 7 \right)_{9}\),\(\left( {16} \right)_{10}\),\(\left( {11} \right)_{11}\),\(\left( {18} \right)_{12}\),\(\left( {20} \right)_{13}\),\(\left( {22} \right)_{14}\),\(\left( {16} \right)_{15}\),\(\left( {11} \right)_{16}\), \(\left( 6 \right)_{17}\),\(\left( 3 \right)_{18}\),\(\left( 8 \right)_{19}\),\(\left( 2 \right)_{20} ]\).

6.1.2 Quantification

The LP model is formulated for cost minimization and is constrained to the required throughput. The cost factors are defined as fixed and operational costs of design-objects and space cost. The cost-related data are defined as attributes of design-objects. ALDeP is applied on alternatives to develop a layout according to the employed design-objects. For more details regarding quantification see Online Appendix A [31]. The quantification results for the demonstrated alternatives are given in Table 2.

Table 2 Warehouse case study result

6.1.3 Fuzzy assessment

The fuzzy framework assesses the alternatives against the qualitative objectives: ‘safety-satisfaction’ and ‘future-expansion’. The importance of objectives was defined by the managers and modelled as fuzzy inputs for the fuzzy models. The objectives’ importance and processes’ importance include four fuzzy sets: ‘Less-Important’, ‘Marginally-Important’, ‘Important’, ‘Very-Important’. Fuzzifications of fuzzy variables are given below:

‘Not-Good’, ‘Not-Efficient’, and ‘less-Important’:

$$ {\text{f}}\left( {{\text{x}}, - 1,0,0.35} \right) = \left\{ {\begin{array}{*{20}l} 1 \hfill & {x < 0} \hfill \\ {\frac{{0.35 - {\text{x}}}}{{0.35}}} \hfill & {0 \le x \le 0.35} \hfill \\ 0 \hfill & {Otherwise} \hfill \\ \end{array} } \right. $$

‘Marginally-Good’, ‘Marginally-Efficient’, and ‘Marginally-Important’:

$$ {\text{f}}\left( {{\text{x}},0.25,0.4,0.6} \right) = \left\{ {\begin{array}{*{20}l} {\frac{{{\text{x}} - 0.25}}{{0.15}}} \hfill & {0.25 \le x \le 0.4} \hfill \\ {\frac{{0.6 - {\text{x}}}}{{0.2}}} \hfill & {0.4 \le x \le 0.6} \hfill \\ 0 \hfill & {Otherwise} \hfill \\ \end{array} } \right. $$

‘Good’, ‘Efficient’, and ‘Important’:

$$ {\text{f}}\left( {{\text{x}},0.45,0.7,0.8} \right) = \left\{ {\begin{array}{*{20}l} {\frac{{{\text{x}} - 0.45}}{{0.25}}} \hfill & {0.45 \le x \le 0.7} \hfill \\ {\frac{{0.8 - {\text{x}}}}{{0.1}}} \hfill & {0.7 \le x \le 0.8} \hfill \\ 0 \hfill & {Otherwise} \hfill \\ \end{array} } \right. $$

‘Very-Good’, ‘Very-Efficient’, and ‘Very-Important’:

$$ {\text{f}}\left( {{\text{x}},0.75,1,1} \right) = \left\{ {\begin{array}{*{20}c} {0 x \le 0.75 } \\ {\frac{{{\text{x}} - 0.75}}{0.25} 0.75 \le x \le 1} \\ { 1 1 \le x} \\ \end{array} } \right. $$

Warehouse managers defined the importance of safety-satisfaction and future-expansion respectively: Important-0.6 and Marginally-Important-0.4. Managers expressed their opinion about efficiencies of design-objects, their interactions, and process importance, which are fuzzy inputs to the fuzzy framework. Details are given in Online Appendix B.

Fuzzy assessment and simulation results of validated alternatives (after applying the feedback algorithm) are shown in Table 2.

6.1.4 TOPSIS

Two criteria were defined for the TOPSIS to assess alternatives: total validated cost and fuzzy assessment result with the relative importance of respectively 0.7 and 0.3. Table 2 shows calculated ranks by TOPSIS for each alternative.

6.1.5 Discussion on case study

\({\text{Alternative}}_{{\text{A}}}\) shows a lower total cost compared to \({\text{Alternative}}_{{\text{C}}}\) according to the quantitation models. These alternatives differ in their employed design-objects in storing/picking processes. \({\text{Alternative}}_{{\text{A}}}\) employs ‘pallet-rack’ and ‘forklift’ in storing and ‘turret-truck’ in the picking process. \({\text{Alternative}}_{{\text{c}}}\) employs ‘shelve-rack’ and ‘forklift’ in both processes. Although, ‘pallet-rack’ is more expensive than ‘shelve-rack’, storing in or picking from the ‘pallet-rack’ takes less time. The ‘turret-truck’ has a shorter operation time compared to the ‘forklift’, but the ‘turret-tuck’ is more expensive. The quantification models calculated that three ‘forklifts’ are required to perform the storing process on ‘pallet-rack’ to satisfy throughput, while four ‘forklifts’ are required if the ‘shelve-rack’ is employed. Three ‘turret-truck’ can satisfy the throughput in the picking process if the ‘pallet-rack’ is employed, while if ‘forklift’ is employed four is needed. Thus, \({\text{Alternative}}_{{\text{A}}}\) demonstrated a lower total cost calculated by quantification models: employing three ‘forklifts’ in storing and three ‘turret truck’ in the picking process.

\({\text{Alternative}}_{{\text{c}}}\) employs ‘CBS’ and ‘double-command’ as its operational policies in storing/picking process. ‘Double-command’ policy allows picking the ordered SKUs after storing some SKUs. Thus, the operator-equipment can cut one empty travelling activity in storing and picking processes. The combination of ‘CBS’ and ‘double-command’ policies reduces the total operation time. In \({\text{Alternative}}_{{\text{c}}}\), the utilization of equipment in storing and picking processes was respectively 59% and 67%. This shows that the total working hours of four ‘forklift’s in \({\text{Alternative}}_{{\text{c}}}\) was between the total working hours of two and three ‘forklifts’ within the available working-time. Therefore, the feedback algorithm reduced the number of ‘forklifts’ in both picking/storing processes to three in the quantification model of \({\text{Alternative}}_{{\text{c}}}\). The quantification models are called again to check the constraints satisfaction in other processes after such an update. Simulation of updated \({\text{Alternative}}_{{\text{c}}}\) revealed the throughput satisfaction.

\(2 \times 5\) < Simulated working-hours for ‘forklift’ < \(3 \times 5\).

 → \(50{\text{\% }}\left( {\frac{2 \times 5 \times 100}{{4 \times 5}}} \right)\) < Utilization < \(75{\text{\% }}\left( {\frac{3 \times 5 \times 100}{{4 \times 5}}} \right)\).

If a sequential design approach was followed without having an integrated design approach in addressing all design requirements at the same time including the operational policies, the results could have been different in cost minimization. It is hard to achieve such insight without having an integrated design approach as realized in this framework.

Both \({\text{Alternative}}_{{\text{A}}}\) and \({\text{Alternative}}_{{\text{c}}}\) received a lower fuzzy score compared to \({\text{Alternative}}_{{\text{B}}}\). This demonstrates the necessity of having a multi-aspect assessment approach. One alternative could be considered promising from the cost minimization perspective, while might not be the best with respect to qualitative objectives. Application of TOPSIS allows ranking the alternatives by considering their excellence on multiple conflicting objectives.

The novel introduced fuzzy framework has a hierarchical structure helping to reduce the number of if–then rules in fuzzy-assessment models by reducing the number of input-variables in each fuzzy-model. The number of if–then rules increases exponentially when the number of variables increases, as called ‘curse of dimensionality’ [59]. The hierarchical structure of the framework helps to break the ‘curse of dimensionality’ in design assessment of CEPSs given the scale and complexity of their design knowledge [60].

6.1.6 Lessons learnt from the case study

Simulation of all alternatives can be timely; hence this research suggests applying a supervisory approach to select and simulate a limited number of alternatives without losing an optimum alternative. However, it is hard to guarantee to find a global optimum because the analytical formulation of CEPS belongs to the NP-hard problem class.

If two subsequent processes do not share resources, if the first process operates properly such that the second can start its operation on time, the employed design-objects in the first process might not impact the operation of the latter process. If the unloading (\({\text{P}}_{11}\)) and the stacking processes (\({\text{P}}_{12}\)) finish their jobs within the available time, their employed design-objects do not impact operations of the storing (\({\text{P}}_{3}\)), shipping (\({\text{P}}_{4}\)), and picking (\({\text{P}}_{3}\)). In this case study, the unloading can employ different design-objects, leading to three possible configurations as shown below. The numbers in parentheses show the UKs of the employed design-objects.

\({\text{P}}_{11} - {\text{Configuration}}_{1}\):[\(\left( 1 \right)_{1}\),\(\left( 3 \right)_{2}\),\(\left( 5 \right)_{3}\),\(\left( 7 \right)_{4}\)].

\({\text{P}}_{11} - {\text{Configuration}}_{2}\):[\(\left( 1 \right)_{1}\),\(\left( 3 \right)_{2}\),\(\left( 6 \right)_{3}\),\(\left( 7 \right)_{4}\)].

\({\text{P}}_{11} - {\text{Configuration}}_{3}\):\([\left( 2 \right)_{1}\),\(\left( 3 \right)_{2}\),\(\left( 6 \right)_{3}\),\(\left( 8 \right)_{4}\)].

There are four feasible configurations for the stacking process, so the number of possible combinations for unloading and stacking process configurations is 3 × 5 = 15. Likewise, the process configurations of storing, picking, shipping lead to 320 combinations.

In total 4800 feasible alternatives are generated. This research selected 15 alternatives that covered all 15 combinations for the unloading/stacking process configurations. The selected 15 alternatives were simulated and validated particularly against their unloading/stacking processes. Then the simulation results of these two processes were written for all feasible 4800 alternatives according to their combination for unloading/stacking process configurations.

Among 15 combinations of unloading/stacking, the combination that had minimum cost while satisfying the throughput was selected. This combination was used to simulate all 320 possible combinations of \([{\text{P}}_{2} ,{\text{P}}_{3} ,{\text{P}}_{4} ]\). However, the simulation results were only updated for storing, picking, and shipping processes. This supervisory approach was done to reduce the computational cost related to the simulation time, so instead of simulating 4800 alternatives, 15 + 320 alternatives were simulated.

6.2 Reconfigurable manufacturing system design case

To illustrate the advantages of the proposed framework compared to the existing approaches, the framework is applied to another case study solved with other approaches in the literature [24, 61]. The used input data are from the literature belonging to the design knowledge of the case study.

This case study is briefly explained here and interested readers are referred to the literature and Online Appendix C and D for details [61]. The case is about designing a reconfigurable manufacturing system to produce two products: ANC-90 and ANC-101. The required hourly throughput for each product respectively is 120 and 180. The objective is cost minimization. Products require common and specific machining operations. Different machine tools are available with different capacities, operation times, and fixed costs. The system is to be designed as a flow line including a maximum of ten sequential and five parallel machines.

Youssef and ElMaraghy [61] used genetic algorithm for cost minimization by finding optimum configuration in terms of the number of stages, machines’ selection, and machines’ number in each stage. Figure 7 shows a found optimal configuration (called \({\text{Altenative}}_{1}\)) with a total cost of $14,645,000. Boxes show the machines’ names.

Fig. 7
figure 7

\({\text{Altenative}}_{1}\)-A possible optimal configuration

Wang and Dagli [24] used a combination of a genetic algorithm and simulation by PN, in which the encoded chromosome took machines’ cost and simulated throughput to choose a new configuration (population). They obtained another optimal configuration (see Fig. 8) with a higher total cost ($17,115,000), called \({\text{Alternative}}_{2}\). They argued that \({\text{Altenative}}_{1}\) could not satisfy throughput by simulation, which this current paper also found a similar issue.

Fig. 8
figure 8

\({\text{Altenative}}_{2}\)-possible optimal configuration

In this case study, the only enabler was defined as machines. This paper defines operational policies as another enabler to show the importance of having an integrated design approach. Accordingly, two possible policies are defined: 'Batching' and 'No-batching'. The described case was aligned with the 'No-batching' policy, in which AN90 and ANC101 could enter the production line one by one. Under the 'Batching' policy, each item can enter the production line in batches consisting of the same item type. \({\text{Alternative}}_{1}^{*}\) and \({\text{Alternative}}_{2}^{*}\) are defined similar to respectively \({\text{Alternative}}_{1}\) and \({\text{Alternative}}_{2}\), except new alternatives use batching policy.

The alternative generation and feasibility checking algorithms developed feasible alternatives, which are quantified by an LP model. The alternatives are simulated for 540 working minutes. The feedback algorithm is applied to alternatives while outputs of the first 60 min are excluded to eliminate the ramp-up period effect. Table 3 shows the results. TOPSIS is applied to alternatives by considering the total cost and throughput with the relative importance of 0.7 and 0.3.

Table 3 Results of second case study

6.2.1 Discussion on case study

The LP results show that \({\text{Alternative}}_{1}\) have a minimum total cost. The simulation results show that \({\text{Alternative}}_{1}\) and \({\text{Alternative}}_{1}^{*}\) cannot produce 300 parts per hour, while \({\text{Alternative}}_{2}\) and \({\text{Alternative}}_{2}^{*}\) satisfy the required throughput. The reason for such a deviation is that different operation times of different machines causes desynchronization between processes and finished parts in one machine cannot immediately get service from the next, causing a queue between stations (bottleneck). Consequently, the machine after bottleneck might work below its nominal capacity, which is against the optimization formulation or assumption. Wang and Dagli [24] argued similarly about the reason for this deviation. The feedback algorithm identified the second process as a bottleneck in \({\text{Alternative}}_{1}\) and \({\text{Alternative}}_{1}^{*}\). Thus, the quantitative models of the alternatives were updated and the number of machines in the second process was increased from ‘one’ to ‘two’.

The modified \({\text{Alternative}}_{1}\) and \({\text{Alternative}}_{1}^{*}\) are simulated again, while \({\text{Alternative}}_{1}\) could not meet the required throughput and \({\text{Alternative}}_{1}^{*}\) could. The feedback algorithm found process-eight as the bottleneck, so the number of machines is increased in the process-eight from ‘two’ to ‘three’. The modified alternative is simulated again and could meet the throughput. \({\text{Alternative}}_{1}^{*}\) shows a better performance compared to \({\text{Alternative}}_{1}\) due to the selection of ‘bathing’ policy (less desynchronization).

The proposed framework can find a better solution in satisfying the objectives compared to found solutions in the literature. The scale of this improvement could be higher if the case study had more realistic assumptions. For instance, if changeover times were addressed, the impact of operational policy selection could be more significant. Moreover, the framework allows ranking the alternatives by considering their performance against qualitative/quantitative objectives. However, the reviewed approaches only considered quantitative ones.

7 Conclusion and future work

Generally, simulation-based optimisation problems face several challenges, which the computational cost is a widely accepted one. The simulation takes a long time when the number of variables is large, while the number of design variables is large at early design stages. Thus, the computational cost may limit the application of simulation-based optimization in this context. Although the simulation-based optimizations do not guarantee finding a global optimum, an investigation of the application of simulation-based optimization can be future work for this work. This research also faced the computational cost problem, which was tried to be addressed with a supervisory approach. A future research direction would be the investigation of the application of more advanced supervisory methods for searching solution space more efficiently.

The SSMM simulates the alternative behaviour under some assumptions (e.g., deterministic operation time of machines). However, the occurrence of some events that are against earlier assumptions may impact the system performance. This is called uncertainty, which can be embodied in SSMM by defining some time-events or probability functions to model possible failures or processing times variability. Accordingly, the simulation results include the impact of uncertainty on an alternative’s performance.

The integrated design of CEPSs with a holistic approach allows capturing more dimensions of the solution space. As shown in Fig. 9, the optimization models seek to find an optimum solution in a solution space whose dimensions is defined by (limited to) quantifiable design requirements. When the solution space is searched by adding the operational policies dimension, a better solution might be identified. Although some limited aspects of operational policies can be modelled with complicated optimization formulation, the possibility of trapping in a local optimum might increase by adding more variables to the optimization models. This highlights the need for using simulation at early design stages.

Fig. 9
figure 9

Multi-dimensional solution space

SSMM is a key artefact in the proposed framework and embodies three aspects of a CEPS: system architecture, decisions regarding the allocation of design options to design requirements for configuring a design alternative, and dynamic behaviour of an alternative. Such a rich embodiment is one of the key contributions of this research and is the fruit of a novel application of proper modelling approach (OO), innovative application of a modelling formalism (FSM), and seamless incorporation of the framework artefacts such as SSMM, fuzzy assessment framework, TOPSIS, quantification models, and supporting algorithms.

SSMM has a hierarchal structure, which its nested states can be used as design modules for encapsulating the detailed design. Therefore, the SSMM can stand as a design platform for designers in different disciplines. The detailed design can solidify the simulation result during design iterations.

Systematic interconnection between models allows a smooth information propagation and realisation of data integrity/consistency. This is the result of the consistent application of architecting guidelines in different models and artefacts of the framework. Such integrity is a significant relief in collaborative design where the design knowledge is large scale and multi-disciplinary.

Because of having a multi-aspect design assessment, this framework allows obtaining a deep insight into the performance of alternatives and the existing trade-offs in objectives’ satisfaction. Therefore, the introduced approach can stand as a decision support framework for the integrated design of complex engineering processed based systems at their early design stages.