1 Introduction

Conventional laser beam welding has certain advantages such as high welding speed, minimal distortion by heat and deep weld penetration. But on combining it with conventional metal inert gas (MIG) welding, the process will be able to bridge large gap and to weld reflective materials with ease. That method is known as Hybrid laser-MIG welding where laser beam and MIG are simultaneously employed in the same weld zone. Nowadays, automotive and shipping industries are increasingly employing this method [1].

Hybrid laser welding is influenced by a large number of controllable process parameters bearing complex relationship among them. Initially, influence of those process parameters on weld bead characteristics [2, 3], microstructure and mechanical properties [2, 4,5,6,7,8], welding strength [9] and welding defects [10] were studied by the researchers. Gradually, optimisation of weld penetration depth [11] and arc stability [12] has been conducted using design of experiment (DOE) based techniques. Zhan et al. [13] determined optimised combination of welding parameters during laser welding of 3-mm and 5-mm-thick TA15 alloy plates using fiber laser. Laser power, welding velocity, and defocus distance were considered as input parameters while heat input, fusion depth, and welding width were considered as measurable output parameters during study. Wang et al. [14] has employed disk laser beam welding technique to fabricate 2-mm 2A14-T6 aluminum alloy plates and found that, finest microstructures, minimum porosity ratio, highest hardness and maximum tensile strength can be obtained when laser power is 2500 W, welding velocity is 2.0 m/min and heat input is 75 kJ/m.

However, for modelling of such multivariable complex process, artificial neural networks (ANN) can be employed as an effective tool. ANN is already known as ‘universal function approximator’ [15] for its ability to model underlying function in a dataset to any arbitrary degree of accuracy. It has been employed for process modelling in electric discharge machining [16], laser cutting [17, 18], laser welding [19] and conventional welding [20]. But all research work mentioned above has employed single hidden layer back-propagation neural network (BPNN) technique [21] for modelling. It is most popular among different feed forward ANN algorithms. Dong et al. [22] optimised connection weights of backpropagation neural networks using genetic algorithm to achieve improved prediction capability and employed for welded joints. Prediction error of less than 5% indicates efficacy of the model developed.

Literatures indicate application of BPNN with gradient descent momentum and BPNN with Levenberg–Marquardt (LM) algorithm for ANN modelling. BPNN with LM is decade faster [23] algorithm compared to traditional BPNN, but sometimes yield poor prediction capability while models noisy dataset. However, a network trained through BPNN with Bayesian regularisation (BR) [24] can perform exceptionally well during testing or prediction with small and noisy dataset. Very few applications of it has been found in literatures. Recently, Chaki and Ghosal [25, 26] have employed it for evolutionary computation-based optimization of hybrid laser welding process. Chaki and Ghosal [25] optimized depth of penetration during CO2 laser–MIG hybrid welding of 5005 Al–Mg alloy using ANN-GA hybrid model where maximum penetration depth of 3.84 mm has been obtained during optimisation with mean absolute % errors of 0.7198%. Chaki et al. [26] further employed a comparative study of ANN-GA, ANN-SA and ANN-Quasi Newton models for optimization of hybrid CO2 laser-MIG process and ANN-GA model has shown best optimization performance with absolute % error of only 0.0503% during experimental validation. In both the works BPNN with Bayesian regularisation (BR) has been used for computing objective function during optimization.

Radial basis function network (RBFN) [27, 28] is another feed forward network suitable for process modeling and already applied for several engineering applications [29, 30]. But its application is yet to be found in laser welding processes.

Literature survey indicates very few works on application of ANN for modelling of hybrid laser welding process parameters. Only BPNN with BR [25, 26] has been used for prediction of process parameters as a part of evolutionary computation-based optimization of hybrid laser welding process. Efficacy of no other ANN training algorithms have been tested so far for prediction of hybrid laser welding processes. However, a detailed comparative study on performance analysis of different ANN training algorithms with various network architecture is required to select a suitable ANN for a process. But no such study has been found to be conducted for hybrid laser welding process in literature.

In the present work, altogether 65 numbers of different networks have been trained and tested with six different ANN training algorithms such as Gradient descent BPNN with momentum, Gradient descent BPNN with momentum and variable learning rate, BFGS quasi Newton BPNN, BPNN with LM algorithm, BPNN with BR and RBFN for CO2 laser–MIG hybrid welding of aluminium alloy sheet. It can be noted that, apart from BPNN with BR, no other ANN algorithms have been trained and tested earlier for laser hybrid welding process. Moreover, to facilitate training and testing of such large number of ANN architecture, a single program has been developed in MATLAB2017a environment that can facilitate training and testing of all 55 different architecture of BPNN using 5 different ANN training algorithms in a single run. Training and testing performance of each network architecture is stored in a file for further comparative study. Literature does not show application of such programs for continuous evaluation of different ANN models and architecture. That program will be useful for similar ANN based process parameters prediction of other manufacturing processes also. However, 10 architecture of RBFN have been computed separately. Dataset used in the modeling has been obtained from a DOE based experimentation on CO2 laser–MIG hybrid welding of aluminium alloy sheet [26] where tensile strength of welded joint has been considered as output quality characteristic. Best ANN has been determined based on the level of accuracy achieved during prediction of output characteristic. Finally, a sensitivity analysis has been carried out to determine effect of process parameters on output. Selected ANN model is capable of determining the required process parameters to achieve a specific welding strength and can be used for online process monitoring and control.

2 Prediction modelling of hybrid laser welding process

In present work ANN has been employed for prediction modelling of hybrid laser welding process. The modelling has been done using the following steps:

  1. (1)

    Generation of experimental dataset

  2. (2)

    Training with Backpropagation neural networks

  3. (3)

    Training with radial basis neural networks

  4. (4)

    Testing of trained ANN

2.1 Generation of experimental dataset

Experimental dataset used for the present work has been taken from a published literature of Chaki et al. [26]. A three level three factor full factorial experimentation was conducted for hybrid laser welding of AA8011 aluminium alloy using a hybrid setup developed in combination of a 3.5 kW CO2 laser welding system (Rofin Slab: CO2 laser) and a MIG welding machine as shown in Fig. 1.

Fig. 1
figure 1

a Experimental Setup and b schematic diagram of hybrid laser welding process

Controllable input parameters during experimentation was considered as (1) Laser Power (P) in kW, (2) welding Speed (Vw) in m/min and (3) Wire Feed Rate (FR) in m/min. The other controllable parameters are kept fixed for the present experiment. The distance between laser beam and arc is kept 2 mm. Focal plane of the laser is kept 1 mm below the job surface. Stand-Off distance of the welding torch is 12 mm from the work piece and the torch angle is maintained at 53° with the job surface. A mixture of Helium and Argon in equal proportion is used as shielding gas with an operating pressure of 2.5 bar.

Hindalco (India) made cold rolled AA8011 grade Aluminium alloy (0.6–1.00% Fe, 0.50–0.90% Si, 0.05% Mg, 0.20% Mn, 0.10% Cu, 0.20% Zn, 0.08% Ti, 0.05% Cr and rest Al) plate of thickness 3 mm has been used as job specimen. In order to perform the butt-welding experiments, 100 mm × 75 mm specimens have been cut from a sheet of aluminium alloy by a machining process. A square butt joint configuration was prepared to fabricate joints. The filler material used for MIG welding has been ER4043 that contains 5.5% Si, around 0.4% other alloying materials (0.007% Mg, 0.17% Fe, 0.08% Cu, 0.005% Mn, 0.04% Ti, 0.0016% Be, 0.02% Sr) and rest Al. The feed wire diameter is 3.15 mm.

An exhaustive set of pilot experiments has been conducted by varying one of the process parameters at a time, while keeping the rest of them at constant value to determine feasible range of input process parameters for the weld seam with smooth appearance and the absence of any visible defects. In this way, the feasible operating regions for the process parameters are selected. Range of operations for process parameters are 2–3 KW for laser power (P), 2–4 m/min for welding speed (Vw) and 2–8 m/min for wire feed rate (FR). Three levels of each input process parameters have been selected for the experiment. Range of controllable factors during experiment along with their levels are given in Table 1. 27 numbers of experiments have been designed and conducted based on 3 factor 3 level full factorial experimental design without replication.

Table 1 Levels of experimental input parameters

Welding strength (WS) of the joint is considered as output parameter and it has been measured from tensile tests carried out in 8801 microprocessor-controlled 10 Ton INSTRON universal tester with an accuracy of ± 0.4%.

The parameters laser power (P), welding speed (Vw) and wire feed rate (FR) are considered as input to ANN architecture and consists input layer. Welding Strength (WS) in MPa is constitute the output layer of ANN. Values of welding strength with respect to different input parameters has been given in Fig. 2. Replication has been carefully avoided during experimentation as it may lead to erroneous ANN training. 80% of experimental dataset randomly considered as training data and rest 20% has been considered as testing data. Input vector (X) and output vector (D) for an ANN architecture is therefore given by, \(\varvec{X} = \left[ {\begin{array}{*{20}c} P & {V_{w} } & {\text{FR}} \\ \end{array} } \right]\), \(\varvec{D} = \left[ {\text{WS}} \right]\). To ensure better training performance, experimental input and output dataset has been normalised between 0 and 1 and denoted as Xnor and Dnor before employing for ANN training and testing operation. Computer Programming for ANN training and testing process is conducted through Neural Network Toolbox of MATLAB R2017a using Intel Core i3-6006U, 2 GHz and 8 GB PC.

Fig. 2
figure 2

Welding strength with respect to different combination of input parameters during full factorial experimentation

2.2 Training with backpropagation neural networks

Backpropagation is an ANN training method, that approximates functional relationship among input and output vectors during training phase by changing the weights associated to a multi-layered feedforward network with differentiable activation function units. Single hidden layer backpropagation neural networks (BPNN) used for the present work consists of three nodes in input layer representing V, E and FR respectively while output layer has one neuron that represents WS. Number of the hidden layer neurons are considered as variable parameter during the study and varied between 5 and 15. Activation functions associated to the hidden layer and output layer neurons are considered as sigmoidal and linear respectively. In back propagation is all neurons are interconnected to each other by network weights. The initial value of the weights is generated randomly and is updated during iterations. In every layer of network bias act exactly as weight on a connection from a unit where activation is always 1. A schematic diagram of backpropagation neural network architecture is given in Fig. 3. BPNN converts weighted sum of input signal (X) through activation function into hidden layer input which is further converted in similar process into output (O) of output layer. Difference between computed output (O) and experimental output (D) is denoted as error. Being a gradient descent method, it minimises total squared error computed by the net and is given by, mean square error,

$$\varvec{MSE} = \frac{1}{Q}\mathop \sum \limits_{i = 1}^{Q} \left( {\varvec{D}_{{{\text{nor}}\,i}} - \varvec{O}_{i} } \right)^{2}$$
(1)

where Q is total number of training data.

Fig. 3
figure 3

Schematic diagram of backpropagation neural network architecture

That MSE is minimised during training process by subsequent updating of weights during iteration satisfies convergence criteria training stops. In order to improve accuracy in prediction and computational speed, different methodology for weight updating has been adopted. In the present work, five different back propagation training algorithms have been used such as gradient descent back propagation with momentum (traingdm), gradient descent back propagation with momentum and adaptive learning rate (traingdx), BFGS quasi Newton backpropagation (trainbfg), backpropagation with Levenberg–Marquardt algorithm (trainlm), and backpropagation with Bayesian regularisation (trainbr).

In the present work, to facilitate training and testing of a large number of ANN architecture using different ANN training algorithms, a separate program has been developed in MATLAB2017a that completes training and testing of all 55 BPNN architecture in a single run and stores the results obtained from training and testing of each network for comparative study. The main program is associated with separate subroutines for different ANN training algorithms. Initially, one subroutine containing a specific training algorithm is called by main program and different ANN architecture are obtained sequentially by increment in numbers of hidden layer neurons by 1 for subsequent training and testing operations. When maximum number of hidden layer neurons is reached for a specific training algorithm, main program calls next subroutine for similar operations using another ANN training algorithm and the similar process as explained continues. Finally, when all subroutines or training algorithms are called and implemented for operations the program terminates. A schematic diagram of the program is given in Fig. 4.

Fig. 4
figure 4

A schematic diagram of the program for backpropagation training and testing process

In gradient descent back propagation with momentum method (traingdm), inclusion of the momentum term accelerates the convergence of backpropagation by allowing the network to respond not only to the local error gradient, but also in the direction of the combination of the current gradient and the previous direction for which the weight corrections have been made. Performance of gradient descent back propagation algorithm is very much sensitive to proper setting of learning rate. Gradient descent back propagation with momentum and adaptive learning rate (traingdx), instead of continuing with constant learning rate throughout training process, increases learning rate if new error term calculated during current iteration is less than old error term or vice versa. That improves rate of convergence.

Quasi newton method approximate Hessian matrix during computation of second order derivative of conventional Newton method of optimisation. BFGS quasi Newton backpropagation (trainbfg) updates the weights using Broyden, Fletcher, Goldfarb, and Shanno (BFGS) formula and results in faster convergence or minimisation of training error. Backpropagation with Levenberg–Marquardt algorithm (trainlm) is particularly a faster training algorithm with high training accuracy as it updates the weights up to second order derivative of training error. It is faster as it avoids complex computation of second order derivatives by approximation of Hessian matrix. Backpropagation with Bayesian regularisation (trainbr) minimises linear combination of sum of the squared errors (SSE) and sum of the squared weights (SSW) instead of minimising errors only. It reduces size of the network and improves accuracy particularly for modelling of small dataset.

2.3 Training with radial basis neural networks

Radial basis neural networks (RBFN) does not bear any weighted connection between input and hidden layer. It computes hidden neuron activations using Gaussian basis function which is an exponential of the Euclidean distance measure between the input vector and a prototype vector that characterizes the signal function at hidden neuron. It can be given by,

$$\varphi \left( X \right) = exp\left( { - \frac{{\left\| {X - \mu^{2} } \right\|}}{{2\sigma^{2} }}} \right)$$
(2)

where μ is the center of the basis function, which is the training data points in this case and σ is the spread factor having a direct effect on the smoothness of the interpolating function. In RBFN with exact interpolator, number of hidden layer neurons are equal to number of training data. As in the present problem there are 22 number of training data (80% of experimental data), number of hidden layers with gaussian activation function is 22. The linearly activated output layer bears a weighted connection with hidden layer and produce output as scalar product of hidden layer output and weight vector. In the present study dataset is fed to network after normalization and spread factor varied from 0.1 to 1.0. Mean square error (MSE) is considered as criterion of performance. A schematic diagram of RBFN architecture is given in Fig. 5.

Fig. 5
figure 5

Schematic diagram of RBFN architecture

2.4 Testing of trained ANN

Training performance of ANN indicates the ease with which a network can identify a known dataset. But efficacy of a trained ANN is dependent on its behaviour in an unknown environment. It is assessed in testing phase through determination of prediction capability, when a trained ANN encounters a set of unknown data (i.e. test data). In the present work, test input dataset (\(\varvec{X}_{\text{nor}}^{\text{Test}}\)) is fed through all networks trained through BPNN and RBFN training algorithm. The resulting ANN predicted output (OTest) is compared with corresponding known experimental test output (\(\varvec{D}_{\text{nor}}^{\text{Test}}\)) to determine the prediction error and is measured by the quantity called testing MSE. It is calculated by,

$${\text{Testing}}\;{\text{MSE}} = \frac{1}{N}\mathop \sum \limits_{i = 1}^{N} \left( {\varvec{D}_{\text{nor i}}^{\text{Test}} - \varvec{O}_{i}^{\text{Test}} } \right)^{2}$$
(3)

where N is total number of test dataset.

3 Performance analysis of ANN prediction modelling

In the present work, 65 ANN architecture have been tested using five different BPNN training algorithm and radial basis function networks. Training functions of MATLAB2017a used for present training and testing, corresponding to those BPNN and RBFN training algorithms are traingdm (gradient descent BPNN with momentum), traingdx (gradient descent BPNN with momentum and adaptive learning rate), trainbfg (BFGS quasi Newton BPNN), trainlm (BPNN with LM algorithm), trainbr (BPNN with BR) and newrbe(radial basis function networks as exact interpolator). During BPNN training different ANN architecture have been achieved by varying hidden layer neurons from 5 to 15. Therefore, each BPNN training algorithm has trained 11 numbers of ANN architecture and altogether 55 numbers of BPNN architecture have been trained and tested in the present work. Best BPNN model has been selected based on prediction performance of ANN during testing of different architecture using BPNN training functions traingdm, traingdx, trainbfg, trainlm and trainbr. The prediction performance of all 55 BPNN networks is presented in Fig. 6 in terms of Testing MSE [Eq. (3)]. It can be clearly observed from Fig. 6 that, overall prediction capability of BPNN with BR is best compared to other training algorithms. Table 2 provides a comparative study of networks with best prediction performance obtained from different training algorithms. It is observed from Table 2 and Fig. 6 that, with 3-6-1 network with Gradient descent BPNN with momentum, 3-12-1 network with Gradient descent BPNN with momentum and variable learning rate, 3-7-1 network with BFGS quasi Newton BPNN, 3-6-1 network using BPNN with LM algorithm and 3-11-1 network using BPNN with BR algorithm provided the best performance for individual algorithm. The first term of the network architecture representation (e.g. 3-6-1, 3-11-1 etc.) indicates number of neurons in the input layer of ANN. Here three number of input layer neurons represents three number of input variables for experimentation such as V, E and FR respectively. The third term of network architecture, represents one neuron in the output layer of ANN which is WS. The middle term of the network architecture indicates number of hidden layer neurons in the ANN architecture.

Fig. 6
figure 6

Testing performance of different BPNN architecture after being trained with different BPNN algorithms

Table 2 Comparison of prediction performance of different ANN models

Figure 7 represents mean absolute % error during prediction along with maximum and minimum error obtained by those best performing networks obtained from training and testing of different BPNN algorithms. Mean absolute % error is obtained as mean of absolute % error in prediction for every test data after corresponding de-normalisation as given below:

$${\text{Absolute}}\;\% \;{\text{Error}}\;{\text{in}}\;{\text{Prediction}} = \left| {\frac{{{\mathbf{D}}_{\text{ i}}^{\text{Test}} - {\mathbf{O}}_{\text{i}}^{\text{Test}} }}{{{\mathbf{D}}_{\text{ i}}^{\text{Test}} }} \times 10 0 { }} \right|$$
(4)

where \({\text{O}}_{\text{i}}^{\text{Test}}\) is output predicted by ANN and \({\text{D}}_{\text{i}}^{\text{Test}}\) represents corresponding experimental output.

Fig. 7
figure 7

Comparison of percentage error in prediction obtained during testing of different ANN models

It has been clearly observed from Fig. 7 that, mean absolute % error, maximum and minimum % error obtained by 3-11-1 network trained using BPNN with BR is least with value of 1.7%, 3.2%, 0.3% respectively. Figure 5 and Table 2 also indicate that, 3-11-1 network trained using BPNN with BR gives the best prediction capability with minimum testing MSE of 3.24E − 04.

Further, 10 architecture of RBFN as exact interpolator has been trained and tested considering spread factor as variable prediction performance of RBFN is presented in Fig. 8. An RBFN with 0.5 spread factor shows best prediction performance with minimum test MSE of 9.56E − 04 as indicated in Fig. 8 and Table 2. The mean absolute % error, maximum and minimum % error obtained by best RBFN network is given by 3.0%, 2.1% and 6.1% respectively which is quite inferior compared to best prediction performance of BPNN obtained by 3-11-1 network trained using BPNN with BR. Therefore, undoubtedly BPNN with BR is considered as the algorithm with best prediction capability for small training dataset as used in the present study and 3-11-1 network trained using BPNN with BR is considered as the best ANN architecture for prediction of welding strength for Laser MIG hybrid welding. A detailed prediction performance of all testing data is finally presented for 3-11-1 network trained using BPNN with BR in Table 3. However, this BPNN with BR training algorithm will be useful for prediction modelling of any small experimental dataset where generating large number of experimental datasets is costly or technically infeasible.

Fig. 8
figure 8

Testing performance of different RBFN architecture

Table 3 Performance of BPNN with BR trained 3-11-1 network during testing with testing input data

4 Regression model of welding strength

A multivariable regression model is developed based on the experimental dataset (Fig. 2) to find out a relationship between three input process variables such as laser power (P), welding speed (Vw) and wire feed rate (FR) with welding strength (WS) as output parameter. Initially, a linear model has been developed but it was later rejected based on the analysis of variance (ANOVA) test results. Finally, the 2nd order regression model is developed for WS and is given below:

$$\begin{aligned} {\text{WS}} & = 246.237 - 118.822 \times P + 15.416 \times V_{w} + 2.65 \times FR + 17.211 \times P^{2} - 5.907 \\ & \quad \times V_{w}^{2} - 0.56 \times {\text{FR}}^{2} + 6.533 \times {\text{P}} \times V_{w} + 1.597 \times P \times {\text{FR}} - 0.729 \times V_{w} \times {\text{FR}} \\ \end{aligned}$$
(5)

It is very important to check the adequacy of the proposed model. Adequacy of the model has been tested by Regression coefficient (R-square) value of the regression equations and ANOVA test. The test results of ANOVA for welding strength is given in Table 4. R2 and R2 (adjusted) for welding strength regression model are 0.9796 and 0.9688 respectively. So the model can be accepted so far as R2 and R2 (adjusted) values are concerned. During ANOVA test, F-value for regression is 90.72 which is much higher than corresponding tabulated F-value, which gives F0.05 (9, 17) = 2.49. Moreover, as P value of regression model is less than 0.05 (alpha at 95% confidence interval), the null hypothesis is rejected. Therefore, regression model for welding strength is considered significant.

Table 4 Results of ANOVA for regression model of welding strength

5 Sensitivity analysis

Sensitivity analysis is a method to identify critical input parameters that exerting the most influence upon model outputs. Mathematically, sensitivity of a design objective function with respect to a design variable is the partial derivative of that function with respect to its variables. Present study is aimed to predict the variation in welding strength due to a small change in process parameters (welding current, voltage and welding speed) for SAW processes. The weld bead characteristic models can be interpreted as design objective functions and their variables as design parameters. To accomplish the need, sensitivity equations [Eqs. (6)–(8)] has been derived from partial differentiation of regression equation mentioned in Eq. (5) with respect to each process parameters (welding current, voltage and welding speed) and are given below:

$$\frac{{{\mathbf{\partial }}({\mathbf{WS}})}}{{{\mathbf{\partial P}}}} = - 118.822 + 34.422 \times {\text{P}} + 6.533 \times {\text{V}}_{\text{w}} + 1.597 \times {\text{F}}_{\text{R}}$$
(6)
$$\frac{{{\mathbf{\partial }}\left( {{\mathbf{WS}}} \right)}}{{{\mathbf{\partial }}{\mathbf{V}}_{{\mathbf{w}}} }} = 15.416 - 11.814 \times V_{w} + 6.533 \times P - 0.729 \times F_{R}$$
(7)
$$\frac{{{\mathbf{\partial }}({\mathbf{WS}})}}{{{\mathbf{\partial FR}}}} = 2.65 - 1.12 \times {\text{FR}} + 1.597 \times {\text{P}} - 0.729 \times {\text{V}}_{\text{w}}$$
(8)

\(\frac{{\partial \left( {WS} \right)}}{\partial P}\), \(\frac{{\partial \left( {WS} \right)}}{{\partial V_{w} }}\) and \(\frac{{\partial \left( {WS} \right)}}{\partial FR}\) have been computed for all experimental dataset and mean of their values corresponding to different levels of process parameters are given in given in Table 5 and Fig. 9. Here, mean of \(\frac{{\partial \left( {WS} \right)}}{\partial P}\), \(\frac{{\partial \left( {WS} \right)}}{{\partial V_{w} }}\) and \(\frac{{\partial \left( {WS} \right)}}{\partial FR}\) indicates sensitivity of P, Vw and FR on welding strength (WS).

Table 5 Mean of sensitivity of process parameters corresponding to their levels
Fig. 9
figure 9

Variation in sensitivity of process parameters to welding strength

Figure 9a indicates high negative sensitivity in low power region which implies decrease in WS with minimum increase in laser power in that region. While it shows high positive sensitivity in high power region. So, maximum welding strength can be observed with high laser power. Overall an increase in WS is observed with increase in laser power. In Fig. 9b high negative value of sensitivity indicates a significant decrease in WS with increase in welding speed. At lower speed low positive sensitivity indicates a rise in WS. In general, WS decreases with increase in welding speed. Figure 9c indicates wire feed rate sensitivity results on WS and trend is negative. As WS is more sensitive to higher wire feed rate, a decrease in WS will be observed with increase in feed rate. Therefore, feed rate should be kept as low as possible to obtain better WS. From Fig. 9a–c it is clear that, wire feed rate is having least sensitivity compared to laser power and welding speed on WS. Therefore, variation of wire feed rate will cause little change of WS and while welding speed and laser power will be most influencing factors. Above analysis indicates that, maximum WS will be obtained if welding is carried out with high laser power, low welding speed and low wire feed rate.

6 Conclusion

In the present work, 65 numbers of different artificial neural network architecture have been employed to estimate tensile strength of hybrid CO2 laser MIG welded aluminium alloy plates. Five backpropagation neural network (BPNN) training algorithms such as (1) gradient descent BPNN with momentum, (2) gradient descent BPNN with momentum and adaptive learning rate, (3) BFGS quasi Newton BPNN, (4) BPNN with LM algorithm and (5) BPNN with BR) and one radial basis function networks (RBFN) training algorithm i)RBFN as exact interpolator have been employed for training and testing of 65 ANN architecture. Further a sensitivity analysis has been carried out to study the influence of process parameters on tensile strength. Finally, following conclusions can be drawn on the basis of results obtained:

  1. (1)

    3-11-1 network during BPNN with BR training and testing results best prediction performance among different architecture with MSE of 3.24E − 04 during prediction of welding strength and is considered as best ANN. Superiority of 3-11-1 network trained using BPNN with BR has been further established as it shows best prediction capability with least value of mean absolute % error, maximum and minimum % error, as 1.7%, 3.2%, 0.3% respectively.

  2. (2)

    Sensitivity analysis of the process indicates that, welding strength increases with increase in laser power and decreases with increase in welding speed. Wire feed rate should be kept as low as possible to obtain better joint strength. However, variation of wire feed rate will cause little change of welding strength and while welding speed and laser power will be most influencing factors. Therefore, maximum welding strength will be obtained if welding is carried out with high laser power, low welding speed and low wire feed rate.

  3. (3)

    A program has been developed in present work for continuous performance evaluation of large number of ANN architecture during training and testing through different ANN training algorithms in a single run. It has been used to determine ANN with best prediction capability for estimation of welding strength during laser MIG hybrid welding process. However, the same program can be efficiently used for estimating output quality of any other processes and can be employed for online process monitoring and control.