Abstract
This paper surveys the artificial neural networks approach. Researchers believe that these networks have the wide range of applicability, they can treat complicated problems as well. The work described here discusses an efficient computational method that can treat complicated problems. The paper intends to introduce an efficient computational method which can be applied to approximate solution of the linear two-dimensional Fredholm integral equation of the second kind. For this aim, a perceptron model based on artificial neural networks is introduced. At first, the unknown bivariate function is replaced by a multilayer perceptron neural net and also a cost function to be minimized is defined. Then a famous learning technique, namely, the steepest descent method, is employed to adjust the parameters (the weights and biases) to optimize their behavior. The article also examines application of the method which turns to be so accurate and efficient. It concludes with a survey of an example in order to investigate the accuracy of the proposed method.
Similar content being viewed by others
Introduction
Recently, integral equations have been extensively investigated theoretically and numerically. Note that they occur in a wide variety of physical applications, various fields of neural sciences and numerous applications such as electrical engineering, economics, elastically, plasticity, etc. Since these equations usually cannot be solved explicitly, it is going to be obtained in approximate solutions. There are several numerical methods for approximating solution of Fredholm and Volterra integral equations in one- and two-dimensions. For example, Tricomi in his book [25] introduced the classical method of successive approximations for integral equations. Variational iteration method [15] was effective and convenient for solving integral equations. The Homotopy analysis method (HAM) was proposed by Liao [16] and then has been applied in [1]. The Taylor expansion approach was presented for solving integral equations by Kanwal and Liu [14] and then has been extended in [17]. In addition, Jafari et al. [12] applied Legendre wavelets method to find numerical solution of linear integral equations. In [13] an architecture of artificial neural networks (NNs) was suggested to approximate solution of linear Fredholm integral equations systems. For this aim, first the truncation of the Taylor expansions for unknown functions was substituted in the origin system. Then the purposed neural network has been applied for adjusting the real coefficients of given expansions in resulting system. In [9], a numerical method based on feed-forward neural networks has been presented for solving Fredholm integral equations of the second kind. The Bernstein polynomials have frequently been applied in the solution of integral equations and approximation theory [5–7, 19, 20]. Also, there are many articles which deal with the solution and analysis of two-dimensional Fredholm and Volterra integral equations. Mirzaei and Dehghan [22] described a numerical scheme based on the moving least squares (MLS) method for solving integral equations in one- and two-dimensional spaces. The method was a meshless method, since it did not require any background interpolation or approximation cells and it did not depend on the geometry of domain. Hadizadeh and Asgary [11] using the bivariate Chebyshev collocation method solved the linear Volterra–Fredholm integral equations of the second kind. Alipanah and Esmaeili [2] approximated the solution of the two-dimensional Fredholm integral equation using Gaussian radial basis function based on Legendre–Gauss–Lobatto nodes and weights. Two-dimensional orthogonal triangular functions are used in [3, 18] as a new set of basis functions to approximate solutions of nonlinear two-dimensional integral equations. Babolian et al. [4] applied two-dimensional rationalized Haar functions for finding the numerical solution of nonlinear second kind two-dimensional integral equations. They reduced the present problem to solve a nonlinear system of algebraic equations using bivariate collocation method and Newton–Cotes nodes. Moreover, some different valid methods for solving these kind of equations have been developed.
This paper focuses on constructing a new algorithm with the use of feed-forward neural networks to reach an approximate solution of the linear two-dimensional Fredholm integral equation. For this purpose, first unknown two-variable function in the problem is replaced by a three-layer perceptron neural network. Supposedly, the limits of integrations are partitioned into set points, this architecture of neural networks can calculate the output corresponding to input vector. Now a cost function to be minimized is defined on the set points. Consequently, the suggested neural net using a learning algorithm that is based on the gradient descent method adjusts parameters (the weights and biases) to any desired degree of accuracy. Here is an outline of the paper. In “Preliminaries”, the basic notations and definitions of the integral equations and the artificial neural networks are briefly presented. “The general method” describes how to find approximate solution of the given two-dimensional integral equations using proposed approach. Finally in “An example”, an numerical example is provided and results are compared with the analytical solutions to demonstrate the validity and applicability of the method.
Preliminaries
In this section we will focus on the basic definitions and introductory concepts in integral equations. In addition the basic principles of artificial neural network (ANN) approach are presented and reviewed for solving linear second kind two-dimensional integral equations (2D-IEs).
Integral equations
Integral equations appear in many scientific and engineering applications, especially when initial value problems for boundary value problems are converted to integral equations. As stated before, we will review some integral equations and linear two-dimensional integral equations of the second kind as well.
Definition 2.1
Let For each partition of and for arbitrary ), suppose
The definite integral of over is
provided that this limit exists in the metric D [25].
Definition 2.2
The linear two-dimensional Fredholm integral equation (2D-FIE) of the second kind is presented by the form [2]
where is a constant parameter, the kernel k and f are given analytic functions on . The two-variable unknown function F that must be determined appears inside and outside the integral signs. This is a characteristic feature of a second kind integral equation. It is important to point out that if the unknown function appears only inside the integral signs, the resulting equation is of first kind.
If the kernel function satisfies in Eq. (1), we obtain the linear two-dimensional Volterra integral equation (2D-VIE) [24]
It should be noted that, if one of the limits of integration varies, the integral equation is called a Volterra–Fredholm integral equation. It is clear that, two-dimensional integral equations appear in many forms. Three distinct ways that depend on the limits of integration are used to characterize these equations which have been are briefly introduced. Notice that, if the function in the present integral equations is identically zero, the equation is called homogeneous. Otherwise it is called inhomogeneous. These three concepts play a major role in the structure of the solution.
Artificial neural networks
Artificial neural networks (ANNs) can be considered as simplified computational structures that are inspired by observed process in natural networks of biological neurons in the brain. They are nonlinear mapping architectures based on the function of the human brain, therefore can be considered as powerful tools for modeling, especially when the underlying data relationship is unknown. A very important feature of these networks is their adaptive nature, where “learning by example” replaces “programming” in solving problems. In other words, in contrast to conventional methods, which are used to perform specific task, most neural networks are more versatile. This feature raises a very appealing computational model which can be applied to solve variety of problems.
The multilayer feed-forward neural network or multilayer perceptron (MLP) that had been proposed by Rosenblatt [23] is very popular and is used more than other neural network type for a wide variety of tasks. The present network learned by back-propagation algorithm is based on supervised procedure. In other words, the network constructs a model based on examples of data with known output.
In this subsection, an architecture of MLP model is discussed here briefly. We intend to give a short review on learning of the given neural network. First consider a three-layer ANN with two input units, N neurons in hidden layer and one output unit. Mathematical representation of the present neural network is given in Fig. 1. Using the figure, input–output relation of each unit and calculated output can be written as follows:
Input units:
The input neurons make no change in their inputs, so:
Hidden units:
Input into a node in hidden layer is a weighted sum of outputs from nodes connected to it. Each unit takes its net input and applies an activation function to it. The input/output relation is normally given as follows:
where describes the result of the net outputs impacting on unit p. Also, are weights connecting neuron i to neuron p and is a bias for neuron p. Bias term is baseline input to a node in absence of any other inputs.
Output unit:
The general method
In this section, we intend to use the MLP method to get a new numerical approach for solving the linear two-dimensional Fredholm integral equation of the second kind. In other words, how to apply this method to make a series approximation for the solution in (1) will be described. The output of two-layer MLP network that is defined in Eq. (1) can be rewritten as follows:
In order to approximate function u, first the intervals and are partitioned into set points and respectively. Thus, the following set of equations will be obtained:
Cost function
First suppose that is the approximate solution with the adjustable parameters (wights and biases) for the unknown After substituting this solution instead of the unknown function in the given 2D-FIE, the Eq. (1) can be transformed to a sum squared error minimization problem corresponding to the proposed neural network. So, the error function is regarded as a function on the weights and biases space of the net for and as follows:
where
Now the total error of the network is defined as:
The goal then is to minimize this function; therefore, we must deduce a back-propagation learning algorithm using the present cost function.
Proposed learning algorithm
Multilayer feed-forward neural network is learned by back-propagation algorithm that is based on supervised procedure. In other words, the MLP network is trained using a supervised learning algorithm which uses the training data to adjust the network weights and biases. Now let and are initialized at small random values for input signals. For parameter adjustment rule can be written as follows:
where r is the number of adjustments, is the learning rate and is the momentum term constant. Similarly this adjustment rule can be written for other weight parameters. Thus, our problem is to calculate the derivative in (11). The derivative can be calculated as follows:
where
Consequently,
Using a similar procedure as mentioned above, we have the correspondingly corollary for parameters and in which we are refrained from going through proof details. So, we have:
and
The MLP neural nets are the sample of regular networks, therefore they can approximate any continuous function on a compact set to arbitrary accuracy [10]. Now the learning algorithm can be summarized as follows:
Learning process
-
Step 1: and are chosen. Then quantities are initialized at small random values.
-
Step 2: Let where r is the number of iterations of the learning algorithm. Then the running error E is set to 0.
-
Step 3: Let Repeat below procedure for different values of i and j:
-
i
Forward calculation: Calculate the output vector by presenting the input vectors and
-
ii
Back propagation: Adjust the parameters using the cost function (8).
-
i
-
Step 4: Cumulative cycle error is computed by adding the present error to E.
-
Step 5: The training cycle is completed. For terminate the training session. If then E is set to 0 and we initiate a new training cycle by going back to Step 3.
An example
In this section, in order to investigate the accuracy of the proposed method, we have chosen an example of linear two-dimensional integral equations of the second kind. For the example, the computed values of the approximate solution are calculated over a number of iterations and the cost function is plotted. Also, to show the efficiency of the present method for our problem, results will be compared with the exact solution.
Example 4.1
Consider the linear 2D-FIE
where
with the exact solution In this example, we illustrate the use of the FNN technique to approximate the solution of this integral equation. In the following simulations, we use the specifications as follows:
-
1.
The number of hidden units:
-
2.
Learning rate
-
3.
Momentum constant
Numerical result can be found in Table 1, and Fig. 2 shows the cost function in the 20 iterations. Figures 3, 4, 5, 6 show the convergence behaviors for computed values of the weight parameters and bias for different number of iterations.
There is no magic formula for selecting the optimum number of hidden neurons. However, some thumb rules are available for calculating number of hidden neurons. A rough approximation can be obtained by the geometric pyramid rule proposed by Masters [21]. For a three-layer network with n input and m output neurons, the hidden layer would have at least neurons.
To show convergence of the proposed method we solve Example 4.1 using shifted Legandre collocation method. The reason for choosing shifted Legandre collocation method is its simplicity. The details of shifted Legandre collocation method are as follows.
Shifted Legandre collocation method
The Legendre polynomials, are the eigenfunctions of the singular Sturm–Liouville problem
Also, they are orthogonal with respect to inner product on the interval with the weight function that is
where is the Kronecker delta. The Legendre polynomials satisfy the recursion relation
where and If is normalized so that then for any n, the Legendre polynomials in terms of power of x are
where denotes the integer part of
The Legendre–Gauss–Lobatto (LGL) collocation points are the roots of together with the points −1 and 1. Explicit formulas for the LGL points are not known. The LGL points have the property that
which is exact for polynomials of degree at most where , are LGL quadrature weights. For more details about Legendre polynomials, see [8].
The shifted Legendre polynomials (ShLP) on the interval are defined by
which are obtained by an affine transformation from the Legendre polynomials. The set of ShLP is a complete -orthogonal system with the weight function Thus, any function can be expanded in terms of ShLP.
The ShLGL (Shifted Legendre–Gauss–Lobatto) collocation points on the interval are obtained by shifting the LGL points, using the transformation
Thanks to the property of the standard LGL quadrature, it follows that for any polynomial p of degree at most on
where , are ShLGL quadrature weights. The results stated above are also satisfied for Legendre–Gauss and Legendre–Gauss–Radau quadrature rules.
The function is approximated by a ShLP of degree at most N as
Now, by substituting (18) and collocation points (17) in (16), we have
By solving this linear system we can find and then approximate the solution
We solved the Example 4.1 using the method described for and Results are shown in Table 2. By comparing Tables 1 and 2 we find that obtained results in Table 1 are in concordance.
Conclusions
This paper suggested a new computational method to solve a two-dimensional Fredholm integral equations. So, a feed-forward artificial neural network has been proposed. This network is able of estimating approximate solution of assumed equation using the learning algorithm which is based on steepest descent rule. Clearly, in order to obtain accurate solution, many learning procedure should be considered. The analyzed examples illustrated the ability and reliability of the present approach. The obtained solutions, in comparison with exact solutions admit a remarkable accuracy. Extensions to the case of more general of integral equations are left for future studies.
References
Abbasbandy, S.: Numerical solution of integral equation: Homotopy perturbation method and Adomians decomposition method. Appl. Math. Comput. 173, 493–500 (2006)
Alipanah, A., Esmaeili, Sh: Numerical solution of the two-dimensional Fredholm integral equations using Gaussian radial basis function. J. Comput. Appl. Math. 235, 5342–5347 (2011)
Babolian, E., Maleknejad, K., Roodaki, M., Almasieh, H.: Two-dimensional triangular functions and their applications to nonlinear 2D Volterra-Fredholm integral equations. Comput. Math. Appl. 60, 1711–1722 (2010)
Babolian, E., Bazm, S., Lima, P.: Numerical solution of nonlinear two-dimensional integral equations using rationalized Haar functions. Commun. Nonlinear Sci. Numer. Simulat. 16, 1164-1175 (2011)
Bhatta, D.D., Bhatti, M.I.: Numerical solution of KdV equation using modied Bernsein polynomials. Appl. Math. Comput. 174, 1255–1268 (2006)
Bhatti, M.I., Bracken, P.: Solutions of dierential equations in a Bernstein polynomial basis. J. Comput. Appl. Math. (2007). doi:10.1016/j.cam2006.05.002
Bhattacharya, S., Mandal, B.N.: Use of Bernstein polynomials in numerical solution of Volterra integral equations. Appl. Math. Sci. 36(2), 1773–1787 (2008)
Canuto, C., Hussaini, M.Y., Quarteroni, A., Zang, T.A.: Spectral methods: fundamentals in single domains. Springer, Berlin (2006)
Effati, S., Buzhabadi, R.: A neural network approach for solving Fredholm integral equations of the second kind. Neural Comput. Appl. doi:10.1007/s00521-010-0489-y
Fuller, R.: Neural Fuzzy Systems. Abo Akademi University, Department of Information Thechnologies (1995)
Hadizadeh, M., Asgary, M.: An ecient numerical approximation for the linear class of mixed integral equations. Appl. Math. Comput. 167, 1090–1100 (2005)
Jafari, H., Hosseinzadeh, H., Mohamadzadeh, S.: Numerical solution of system of linear integral equations by using Legendre wavelets. Int. J.Open Prob. Comput. Math. 5, 63–71 (2010)
Jafarian, A., Measoomy Nia, S.: Utilizing feed-back neural network approach for solving linear Fredholm integral equations system. Appl. Math. Mode. (2012). doi:10.1016/j.apm
Kanwal, R.P., Liu, K.C.: A Taylor expansion approach for solving integral equations. Int. J. Math. Educ. Sci. Technol. 20, 411–414 (1989)
Lan, X.: Variational iteration method for solving integral equations. Comput. Math. Appl. 54, 1071–1078 (2007)
Liao, S.J.: Beyond perturbation: introduction to the homotopy analysis method. Chapman Hall/CRC Press, Boca Raton (2003)
Maleknejad, K., Aghazadeh, N.: Numerical solution of Volterra integral equations of the second kind with convolution kernel by using Taylor-series expansion method. Appl. Math. Comput. 161, 915–922 (2005)
Maleknejad, K., Jafari Behbahani, Z.: Applications of two-dimensional triangular functions for solving nonlinear class of mixed Volterra-Fredholm integral equations. Math. Comput. Mode (2011). doi:10.1016/j.mcm.2011.11.041
Maleknejad, K., Basirat, B., Hashemizadeh, E.: A Bernstein operational matrix approach for solving a system of high order linear Volterra–Fredholm integro-differential equations. Math. Comput. Mode 55, 1363–1372 (2012)
Mandal, B.N., Bhattacharya, S.: Numerical solution of some classes of integral equations using Bernstein polynomials. Appl. Math. Comput. 190, 1707–1716 (2007)
Masters, T.: Practical neural network recipes in C++. Academic press, NewYork (1993)
Mirzaei, D., Dehghan, M.: A meshless based method for solution of integral equations. Appl. Numer. Math. 60, 245–262 (2010)
Rosenblatt, F.: The perceptron: a probabilistic model for information storage ang organization in the brain. Psychol. Rev. 65, 386–408 (1958)
Tari, A., Rahimib, M.Y., Shahmorad, S., Talati, F.: Solving a class of two-dimensional linear and nonlinear Volterra integral equations by the differential transform method. J. Comput. Appl. Math. 228, 70–76 (2009)
Tricomi, F.G.: Integral equations. Dover Publications, New York (1982)
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
This article is published under license to BioMed Central Ltd. Open Access This article is distributed under the terms of the Creative Commons Attribution License which permits any use, distribution, and reproduction in any medium, provided the original author(s) and the source are credited.
About this article
Cite this article
Asady, B., Hakimzadegan, F. & Nazarlue, R. Utilizing artificial neural network approach for solving two-dimensional integral equations. Math Sci 8, 117 (2014). https://doi.org/10.1007/s40096-014-0117-6
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s40096-014-0117-6