Abstract
Deep learning is a wildly popular topic in machine learning and is structured as a series of nonlinear layers that learns various levels of data representations. Deep learning employs numerous layers to represent data abstractions to implement various computer models. Deep learning approaches like generative, discriminative models and model transfer have transformed information processing. This article proposes a comprehensive review of various deep learning algorithms Multi layer perception, Self-organizing map and deep belief networks algorithms. It first briefly introduces historical and recent state-of-the-art reviews with suitable architectures and implementation steps. Moreover, the various applications of those algorithms in various fields such as wireless networks, Adhoc networks, Mobile ad-hoc and vehicular ad-hoc networks, speech recognition engineering, medical applications, natural language processing, material science and remote sensing applications, etc. are classified.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Nowadays, deep learning models significantly impact extracting information or hidden patterns from enormous amounts of data with greater accuracy. Compared to conventional machine learning approaches, deep learning can solve complex problems and correlate the interdependent variables. However, traditional neural networks SOM, MLP and DBN have limited high-level data abstraction, which can be alleviated, when combined with deep neural nets. This paper analyzes different hybrid neural networks MLP, SOM and DBN with deep learning models [1] and the techniques used to improve the performance of the model. MLP, SOM, and DBN models are highly applicable when human experts are unavailable, unable to explain the decisions made using their knowledge adequately, problem solutions evolve and size, and in instances where solutions must be modified depending on new information [2].
2 Multi-Layer Perceptron (MLP)
It is the commonly used neural network based on supervised learning in which information flows in one direction and has no loops. The main objective is to find the optimized function f() that maps input to the desired output and to learn the optimized bias value (θ) for it. Learning occurs in the MLP using a back propagation algorithm by adjusting the connection weights when there is a deviation between the expected and actual output. Their main applications are to solve optimization problems in finance, transportation, fitness, and energy (Figs. 1 and 2).
3 Architecture, Algorithm and Characteristics of MLP
It has three layers: input, output and one or more hidden layers. The input layer collects the input features to be processed. An arbitrary number of hidden layers lies between the input and output layers. They work as the computational unit of the MLP. The output unit performs tasks such as prediction and classification.
-
Property 1: Universality
MLP is capable of learning both linear as well as non-linear functions. MLPs are designed to approximate any continuous function and can solve problems that are not linearly separable.
-
Property 2: Adaptive learning and Optimal
MLP can learn how to do tasks from the data given for training and initial experience. MLP minimizes the loss function. Hence it is optimal. Learning the function that maps the inputs to the outputs reduces the loss to an acceptable level.
-
Property 3: Stochastic
MLP is a stochastic program. In a stochastic program, some or all problem parameters are uncertain and use probability distributions to solve highly complex optimization problems.
-
Property 4: The power of depth
Compared to shallow ones, deep nets can represent some functions more compactly, such as parity function and a deep network, whose size is linear in the number of inputs computes it.
3.1 Application of MLP
There are various convolution neural network-based models for remote sensing image classification and better performance. VHR remote sensing image scene classification plays a vital role in remote sensing research; hence they help manage land resources, urban planning, tracking of disasters, and traffic monitoring. Osama A. Shawky et al. [3] proposed a VHR (Very High Resolution) image scene classification model comprising three phases: Data augmentation to learn robust features, a pre-trained CNN model to extract features from the original image, and an adaptive gradient algorithm multi-layer perceptron to improve the accuracy of the classifier.
With the advent of modern remote sensing technologies, various very fine spatial resolutions (VFSR) dataset is now commercially available. These VFSR images have opened up many opportunities such as urban land use rescue, agriculture, and tree crown description. Zhang et al. [4] proposed a hybrid classification system that combines the contextually based classifier CNN and pixel-based classifier MLP with a rule-based decision fusion strategy. The decision fusion rules formed based on the confidence distribution of the contextual-based CNN classifier. If the input image patch is at the homogeneous region, the confidence is high.
On the other hand, if the image pixel contains other land cover classes as related information, the confidence is low. As a result, the MLP can rectify the classified pixels with low confidence at the pixel level. This paper also compares the proposed method's performance with benchmark standards such as pixel-based MLP, spectral texture-based MLP, and contextual-based CNN classifiers.
Md Manjurul Ahsan et al. [5] proposed a hybrid model with a combination of a Convolutional Neural Network (CNN) and Multilayer Perceptron (MLP) in which MLP handles the numerical/categorical data, and CNN extracts features from the X-ray images. Parameter tuning used the grid search method to decide the number of hidden layers, number of neurons, epochs, and batch size. Meha Desai et al. [6], in their study, compare and analyze the function and designing of MLP and CNN for the application of breast cancer detection and conclude CNN give slightly higher accuracy than MLP.
Vinod Kumar et al. [7] suggested a hybrid CNN-MLP model that analyzes novel and diversified attacks. The problem of intrusion detection is a classification task using machine learning and deep learning techniques. The model used feature selection and reduction techniques, random forest regressor, along the correlation parameter. The CICIDS2017 dataset used the performance of the proposed model outperforms that of the performance of the individual CNN and MLP models. Hanwen Feng et al. [8] suggested a CNN model for Classification of Points of Interest in Side-channel attacks and compared it with MLP and concludes that MLP is more suitable for PCA traces and CNN is for POI traces; shorter traces improves the classification results. Bikku [9] proposed a model using MLP to predict future health risk with a certain probability, comparing it with LSTM and RNN, and they suggested MLP outperforms the other two. Salah, L. B presented a model to control a bioreactor using deep learning feed-forward neural networks with different MLP structures. The trained model emulates the inverse dynamics of the bioreactor and then uses neural controllers for neural control strategies of the chosen bioreactor [10].
Bairavel et al. [11] suggested a model for multimodal sentiment analysis using feature-level fusion technique and novel oppositional grass bee optimization (OGBEE) algorithm for fusing the extracted features from different modalities and MLP for classification. [12] compared three different neural network approaches, MLP, RBF and PNN, for Thematic mapping from remotely sensed data. For the proposed model, PNN outperforms. Singh et al. [13] designed a model to find the optimal collision-free path and control the robot's speed in a dynamic environment for the mobile robots to reach the destination using MLP. The ultrasonic sensors in the robot sense the obstacle in its path and calculate the distance between them. Meng Wang et al. [14] devised a model to detect the Distributed Denial of Service (DDoS) attack using MLP with feature selection for optimal feature selection and the Back Propagation algorithm to reconstruct the detector when errors are perceived. The model comprises three modules knowledge base, detection model, and feedback mechanism and MLP act as binary classifier during attack detection [15] (Table 1).
The model proposed by Morteza Taki et al. [15] predicts the irrigated and rainfed wheat output energy using artificial network models MLP, RBF and Gaussian Process Regression (GPR). The RBF model performs better than the other two models in predicting wheat output energy under various irrigated and rainfed farms. [16]
In order to decrease error and packet loss in the network, Jafari-Marandi et al. [18] suggested an MLP with many layers arranged in an input layer, an output layer, and several hidden layers. Back-propagation learning is a method used by MLPs to train the network. A collection of vectors derived from actual data or produced by a realistic simulator are used during the training phase. Perceptrons learn by adjusting the connection weights to reduce the output error relative to the predicted outcome for each training vector. Authors [19] propose secure routing on the Internet of Vehicles based on V2V (Vehicle to Vehicle) communication [20] using multilayer perceptron (MLP) to detect intruders or attackers on the network [21]. MLP algorithm-generated classification report based on Denial of service attacks. The MLP's precision, recall, F1 score, and support [22] scores are displayed in the classification report. They utilized it to compare classification models and choose the ones with more robust or balanced classification scores. The measures employed are true positives and false negatives as well as true and false positives [23].
4 Self-organizing Map (SOM)
A Self-organizing map (SOM) is an unsupervised-based neural network algorithm. It is also referred to as the dimensionality reduction algorithm or Kohonen Network with input and output layers without a hidden layer. Since this algorithm reduces the input dimension of the data, the final output is represented as a feature map. Similar samples of maps are merged as a map. Generally, this algorithm is used to convert the high-dimensional dataset into a 2D discretized pattern [4].
4.1 Architecture, Algorithm and Characteristics of SOM
The SOM architecture has two layers: the input and output layers with the feature map. It does not contain any hidden layer like a neural network. So, it just passes weight values to the output layer without performing any activation function in neurons. At the same time, each neuron is assigned with some weight value based on the input space. The SOM architecture has a feed-forward structure with a 2D computational layer of nodes arranged in rows and columns and connected fully with all other sources of the input layer [24]. Figure 3 depicts an architectural overview of Kohonen’s SOM.
The SOM uses competitive learning to update its weights. It consists of three methods as Competition, Cooperation and Adaptation. In the competition process, compute the distance between each neuron of the Kohonen layer and the input layer and identify the minimum or maximum distance of neuron based upon the applications. That will be considered as the winner of the process. Following the cooperation process, select the neighbourhood neurons depending on the time and distance of the winner neurons.
At last, the adaptation process updates the weight values of the winner and cooperative neurons. Finally, it produces a feature map from input variables. The main properties of the SOM are described as, Property 1: Approximation of the Input Space: The feature map in the output space, which is expressed by a collection of weight vectors, is a fair estimation of the input space. Property 2: Topological Ordering: The SOM algorithm produces a topologically ordered feature map, meaning that the spatial position of a neuron in the output lattice or grid correlates to a specific domain or feature of the input data. Property 3: Density Matching: The feature map represents differences in the input distribution's statistics regions in the input space. For example, the high probability sample training data are mapped into more significant domains of the output space, and thus with higher resolution, than regions of input space where training vectors are produced with low probability values. Property 4: Feature Selection: The self-organizing map will pick a set of optimal features for quantifying the underlying distribution of given data from input data with a non-linear distribution. Figure 4 represents the algorithmic steps of SOM.
4.2 Application of SOM
Kohonen et al., [62] initially used them for speech recognition. But nowadays it is used in various applications such as, Pattern recognition, speech processing, industrial and medical diagnostics applications [25, 26], and data mining etc. using some hybrid architecture merged with RNN, CNN and back tracking approaches. The phonetic typewriter of Kohonen is one of the earliest and most well-known applications of the SOM. The challenge is to identify phonemes in real time so that they can be used to drive a typewriter from dictation in the field of speech recognition. The speech signals [27] are pre-processed before applied to the SOM. The Fourier transforming and filtering process are used to sample the data using 24 dimensional spectral vectors. The proposed network was effectively trained using speech waveforms, and the output nodes are naturally clustered with the ideal phonemes. Finally, the model output generated logical phoneme strings from real-world speech applications (Table 2)
Behnisch and Ultsch [64] used an extension of SOM as an emergent self-organizing map (ESOM) for clustering and classification of the data. It keeps the high-dimensional data's neighborhood relationships. On the other hand, the finite grid has a drawback because neurons on the map's edges have somewhat different mapping qualities than neurons in the middle versus those on the boundary. Growing hierarchical self-organizing maps (GHSOM) used by Chifu and Letia [65] consist of a set of SOMs (particularly bidimensional grids) organized as nodes in a hierarchy. GHSOM is initialized with a hierarchy mirroring the one in the taxonomy, and concepts are mapped to some nodes in the corresponding SOM by initializing the node’s weights with the vector description of the concept; all other unmapped nodes are initialized randomly [28].
Potok [29] suggested the architecture for Breast cancer diagnosis using neural network based self-organizing maps (SOM). It classified tumour breast or benign and non tumour or malignant lesions with the samples of 243 breast tumours. The SOM based autocorrelation texture features used to classify the tumours. Osman [30] suggested a corona virus detection technique based on the Locality weighted learning and self-organization map (LWL-SOM) technique for capturing the images and identify the diseases of COVID-19 cases. They grouped the chest X-ray data patterns based on SOM strategy to categorize between the positive and negative cases of COVID-19. Then, they built locality weighted learning model for diagnosing the cases. Materials Informatics application.
Jimin et al. [28] used the machine learning technique of SOM to visualize and validate relationships between high-dimensional materials dataset. Compared to conventional methods, the SOM categorized and validated the materials using various mapping techniques like U-matrix map, heat maps, cluster-based map, and Gruneisen parameter. Nicolas et al., [35], proposed the classification of vertebral problems using K-means and SOM algorithm. In his approach, SOM outperformed than the K-Means clustering analysis. Felix et al. [36] and Srivatsa et al., [37] used the hybrid version of self organising map algorithm with Susi framework and Cellular Self-Organizing Map for classifying the hyper spectral datasets. A self-organization-based clustering network in MANET employing zone-based group mobility was proposed by Farooq et al. in [40] to increase scalability and decrease additional energy consumption in the network. They employed the flocking behavior of birds as a microbe model to build and maintain clusters in MANETs. A cluster size management plan that lowers network traffic, enhances MANETs' performance in group mobility, and uses less energy. A distance-based intrusion detection system based on an unsupervised SOM network was proposed by Barletta et al. in [41]. To identify intrusions, the SOM network connects to CAN buses inside the car. Many hybrid methods, combining the SOM network with other clustering methods, such as the k-means algorithm, have been proposed to improve the accuracy of the model. Several data sets, including those for automobile hacking, spoofing, fuzzy data, and denial-of-service attacks, are used for simulation.
5 Deep Belief Network
Nowadays, Machine Learning dominates research interests due to its vast application in various fields. “Deep Learning,” a type of machine learning algorithm also known as Representation Learning [42] has its application in multimedia concept retrieval, text mining, social network analysis and video recommendation. Deep Learning represents ANN with layered network topologies of neuron models (Fig. 5).
5.1 Architecture, Algorithm and Characteristics of DBN
Hinton et al. [43], Deep Belief Network (DBN) [22] is a popular deep learning algorithm representing advanced learning methodology, more profound architecture, and high-level abstraction of biological modelling, giving rise to simplified mathematical models. Its network architecture is inspired by artificial intelligence (AI) research study replicate human-level intelligence.
DBN, an alternate class of Deep Neural Network, is a graphical model with multiple layers of ‘hidden units’ with a connection within layers and not within each layer [22]. Trained Unsupervision DBN reconstructs its inputs probabilistically acting as feature detectors, whereas trained Supervision DBN is utilized for classification. Restricted Boltzmann Machine (RBM)/autoencoders is an unsupervised DBN representing an undirected, generative energy-based model where hidden layer acts as a visible layer for the successor. When trained greedily layer-wise, DBN leads to an effective deep learning algorithm [44]. Overall deep belief network real-time application scenarios include computer vision, electroencephalography [45], drug discovery [46], natural language processing, speech recognition, material inspection, board game programs, and so forth, with fantastic outcomes surpassing human expert achievement. Figure 6 represents the algorithmic steps of the Deep Belief Network.
5.2 Application of DBN
Liu et al. [47] suggested a model for image classification using Deep belief network. In this, the stacked restricted Boltzmann machine make use of contrastive divergence algorithm for feature extraction and softmax layer make use of Evolutionary Gradient Descent(EGD) strategy to classify the extracted features. The acceleration rate of EGD is remarkable compared to Gradient Descent algorithm. Jianjian Yang et al. [48] proposed a model for deep fault recognition using Deep Belief Network. To refine the result of the DBN, the proposed method makes use of stochastic adaptive particle swarm optimization (RSAPSO) algorithm. To address the limitations such as local optimization and low search accuracy of conventional PSO algorithms, the proposed system used RSAPSO algorithm that allows particles to reset the position from the original with an assured probability, and continue its searching again. Hence, the proposed method minimizes the probability of the trapping in a local minimum of particle swarm. Dan Wang et al., [49] presented a system that applied raw physiological data to Deep Belief Networks (DBNs) with three classifiers to envisage the levels of emotions such as arousal, valence, and liking based on the known features [50].
The classification accuracies obtained are better than the results acquired by Gaussian Naïve Bayes classifier. Xiaoai Dai et al., [51] devised a DBF model to extract artificial target features in cities, as a hyper-spectral image. DBF performs dimensionality reduction and extracts the depth features of pixels. DBF provides better robustness and separability compared with Principal component analysis. O’Connor et al., [52] This paper proposes a method based on the Siegert approximation for Integrate-and-Fire neurons to map an offline-trained DBN onto an efficient event-driven spiking neural network suitable for hardware implementation. The method is demonstrated in simulation and by a real-time implementation of a 3-layer network with 2694 neurons used for visual classification of MNIST handwritten digits with input from a 128 × 128 Dynamic Vision Sensor (DVS) silicon retina, and sensoryfusion using additional input from a 64-channel AER-EAR silicon cochlea. The system is implemented through the open-source software in the jAER project and runs in realtime on a laptop computer. It is demonstrated that the system can recognize digits in the presence of distractions, noise, scaling, translation and rotation, and that the degradation of recognition performance by using an event-based approach is less than 1%. Movahedi, faezeh et al., [46] discussed the state of- the-art algorithms for deep belief networks and their performances in electroencephalographic applications in medical fields such as emotion recognition, sleep stage classification, and seizure detection. It also includes the challenges and future research direction of DBF in electroencephalographic applications. Abdellaoui and Douik [53] suggested an optimal HAR system with a two-phase DBN model that offers a better quality of classification prediction [54] (Table 3).
Kuldeep et al. [55], proposed to build a cluster out of the vehicles based on their relay nodes, distrust values, and recommendation and experience-based faith vehicles. Deep Belief Network (DBN) used the threshold value discovered to classify the vehicles into three lists: regular, irregular, and vicious. By removing the malicious nodes from the VANET network [56], this belief result from the FBTRP-DBN model enables the selection of an ideal cluster head using the highest trustworthy node [57]. According to that it improved the overall connectivity and throughput of the network [26, 58]. Moreover, the classification algorithm provided the varied results based on the involving vehicles types and counts i.e.real time data sets [59]. The challenges faced by these algorithms are depicted in the Fig. 7.
5.3 Limitations of MLP, SOM and DBN Algorithms
MLP with back propagation has feeble generalization ability for statistically neutral problems. Hence the model does not know the expected output, and the relation between different input variables determines the output.MLP has too many parameters as it is fully connected, resulting in redundancy and inefficiency [6].MLP disregards Spatial information in solving problems. SOM minimizes the volume of the dataset, making it easier to visualize and form clusters. However, it has several flaws, including poor handling of categorical variables. Solutions for enormous data sets are computationally demanding and potentially inaccurate. As per DBN, it needs High energy and large space requirement for its execution. The huge requirement of Random Number Generators (RNGs) drops deep belief networks energy efficiency [60].
Moreover, in DBN, Learning time is prolonged in a back-propagation neural network with multiple hidden layers. Greedy learning is inefficient in the directed module as the posterior is not factorizable in each training case. Integrating the overall organization of higher variables before the first hidden layer makes learning layer-wise difficult in a sigmoid belief network.MLP has recently seen a sharp decline in research and has been used in many tasks, including content retrieval, intrusion detection, video recommendation, and image classification. SOM and DBN are two of the more often used compared MLP algorithms in these applications [61]. The ongoing emergence of new research in deep and distributed learning results from the ever-increasing ease with which data can be obtained and the incredible advancements in hardware technologies like high-speed computing.
The main challenges of these algorithms are lack of training data in intrusion detection, imbalanced data issues in ad-hoc network applications, Interpretability of data, Uncertainty scaling in health care, Catastrophic forgetting in biological applications, Model compression in health care, Over fitting in medical applications during misbehaviour detection in mobile networks, Exploding Gradient and Vanishing gradient problem in energy efficient network formation and under specification of data in secure routing. Cloud-based platforms are anticipated to be crucial in the future creation of computational Deep learning applications. Handling the massive volume of data can be done with cloud computing. Additionally, it aids with cost- and efficiency-saving measures. Additionally, it gives the option to train Deep learning architectures [61].
6 Conclusion
The field of machine learning simulating deep learning methods has slowly become the dominant domain in this era. This chapter discusses the issues and challenges and aggregates numerous existing solutions in medical, material and NLP applications. Though, there are still several issues and challenges that need to be tackled in the future. MLP is one of the deep learning approaches with universality and stochastic properties. Hence its application is versatile, such as intrusion detection in ad-hoc networks, medical image processing for disease prediction, cluster-based data transformation in VANET, remote sensing for image scene classification and networks to identify the attacks, energy engineering to predict the output energy in various applications, and robot navigation to determine the optimal collision-free path the destination and so on. The self-organizing maps and modified versions of SOM are discussed based on various application platforms. The unsupervised-based learning algorithm has some standard steps and properties to create a map from the large or complicated input data set. Due to its high adaptiveness, many applications like medical diagnosis, data compression techniques, Bibliographic classification, and Image browsing have used this algorithm for classification. Deep learning is the fast-growing approach that provides solutions for hectic challenges in various applications. Moreover, machine learning is transforming into its new phase of intelligent Artificial Intelligence applications .
Availability of data and material (data transparency)
Not Applicable.
Code availability (software application or custom code)
Not Applicable.
References
Alom, M. Z., Taha, T. M., Yakopcic, C., Westberg, S., Sidike, P., Nasrin, M. S., Hasan, M., Van Essen, B. C., Awwal, A. A. S., & Asari, V. K. (2019). A state-of-the-art survey on deep learning theory and architectures. Electronics, 8(3), 292. https://doi.org/10.3390/electronics8030292
Ghasemi, F., Mehridehnavi, A., Pérez-Garrido, A., & Pérez-Sánchez, H. (2018). Neural network and deep-learning algorithms used in QSAR studies: Merits and drawbacks. Drug Discovery Today, 23(10), 1784–1790. https://doi.org/10.1016/j.drudis.2018.06.016
Shawky, O. A., Hagag, A., El-Dahshan, E.-S.A., & Ismail, M. A. (2020). Remote sensing image scene classification using CNN-MLP with data augmentation. Optik. https://doi.org/10.1016/j.ijleo.2020.165356
Li, W., Zhang, S., & He, G. (2013). Semisupervised distance-preserving self-organizing map for machine-defect detection and classification. IEEE Transactions on Instrumentation and Measurement, 62(5), 869–879. https://doi.org/10.1109/TIM.2013.2245180
Ahsan, M. M., Alam, T. E., Trafalis, T., & Huebner, P. (2020). Deep MLP-CNN model using mixed-data to distinguish between COVID-19 and non-COVID-19 patients. Symmetry, 12(9), 1526. https://doi.org/10.3390/sym12091526
Desai, M., & Shah, M. (2021). An anatomization on breast cancer detection and diagnosis employing multi-layer perceptron neural network (MLP) and Convolutional neural network (CNN). Clinical eHealth, 4, 1–11. https://doi.org/10.1016/j.ceh.2020.11.002
Kumar, V., Rana, K., Malik, J., & Tomar, A. (2020). Evaluating hybrid Cnn-Mlp architecture for analyzing novel network traffic attacks. International Journal of Scientific & Technology Research, 9(03), 2277–8616.
Feng, H., Lin, W., Shang, W., Cao, J., & Huang, W. (2020). MLP and CNN-based classification of points of interest in side-channel attacks. International Journal of Networked and Distributed Computing, 8(2), 108–117.
Bikku, T. (2020). Multi-layered deep learning perceptron approach for health risk prediction. Journal of Big Data, 7, 50. https://doi.org/10.1186/s40537-020-00316-7
Salah, L. B., & Fourati, F. (2019). Deep MLP neural network control of bioreactor. In: 2019 10th International Renewable Energy Congress (IREC). https://doi.org/10.1109/irec.2019.8754572
Bairavel, S., & Krishnamurthy, M. (2020). Novel OGBEE-based feature selection and feature-level fusion with MLP neural network for social media multimodal sentiment analysis. Soft Computing, 24, 18431–18445. https://doi.org/10.1007/s00500-020-05049-6
Hasan, R. I., Yusuf, S. M., & Alzubaidi, L. (2020). Review of the state of the art of deep learning for plant diseases: A broad analysis and discussion. Plants, 9(10), 1302.
Singh, N. H., & Thongam, K. (2018). Mobile robot navigation using MLP-BP approaches in dynamic environments. Arabian Journal for Science and Engineering, 43, 8013–8028. https://doi.org/10.1007/s13369-018-3267-2
Wang, M., Lu, Y., & Qin, J. (2020). A dynamic MLP-based DDoS attack detection method using feature selection and feedback. Computers & Security, 88, 101645.
Taki, M., Rohani, A., Soheili-Fard, F., & Abdeshahi, A. (2018). Assessment of energy consumption and modeling of output energy for wheat production by neural network (MLP and RBF) and Gaussian process regression (GPR) models. Journal of Cleaner Production, 172, 3028–3041. https://doi.org/10.1016/j.jclepro.2017.11.107
Chen, D., Chang, R. F., & Huang, Y. L. (2000). Breast cancer diagnosis using self-organizing map for sonography. Ultrasound in Medicine and Biology, 26(3), 405–411. https://doi.org/10.1016/s0301-5629(99)00156-8
Zhang, C., Pan, X., Li, H., Gardiner, A., Sargent, I., Hare, J., & Atkinson, P. M. (2018). A hybrid MLP-CNN classifier for very fine resolution remotely sensed image classification. ISPRS Journal of Photogrammetry and Remote Sensing, 140, 133–144. https://doi.org/10.1016/j.isprsjprs.2017.07.014
El Khatib, A., Mourad, A., Otrok, H., Abdel Wahab, O., & Bentahar, J. (2015). A cooperative detection model based on artificial neural network for VANET QoS-OLSR protocol,https://doi.org/10.1109/ICUWB.2015.7324400
Anzer, A., & Elhadef, M. (2018). A multilayer perceptron-based distributed intrusion detection system for internet of vehicles. In: 2018 IEEE 4th international conference on collaboration and internet computing (CIC).
Naskath, J., Paramasivan, B., Mustafa, Z., et al. (2021). Connectivity analysis of V2V communication with discretionary lane changing approach. Journal of Supercomputing. https://doi.org/10.1007/s11227-021-04086-8
Damodaram & Pavani, K. (2013). Intrusion detection using MLP for MANETs. In: International conference on computational intelligence and information technology (CIIT 2013).
Hinton, S., & Osinder, Y. W. (2006). Teh ‘A fast learning algorithm for deep belief nets.’ Neural Computation, 18(7), 1527–1554.
Abdan, M., & Seno, S. A. H. (2022). Machine learning methods for intrusive detection of wormhole attack in mobile ad hoc network (MANET). Wireless Communications and Mobile Computing, 2022, 12.
Liao, G., Chen, P., Du, L., Su, L., Liu, Z., Tang, Z., et al. (2015). Using SOM neural network for X-ray inspection of missing-bump defects in three-dimensional integration. Microelectronics Reliability, 55, 2826–2832.
JanoferIbrahima, J., Naskath, J., Lakshmi Prabha, S., & Paramasivan, B. (2020). Phone directory using mobile application. International Journal of Scientific and Technology Research, 9(3), 6495–6498.
Muthukaviya, S., Shree, M.R., Vardhini, S.R.V.S., & Naskath, J. (2022). IoT based trolley for isolated patients. In: AIP conference proceedings, 2444 (1).
Estrebou, C., Lanzarini, L., & Hasperuée, W. (2010). Voice recognition based on probabilistic SOM, In: Conferencia Latinoamericana en Informática.
Qian, J., Nguyen, N. P., Oya, Y., Kikugawa, G., Okabe, T., Huang, Y., & Ohuchi, F. S. (2019). Introducing self-organized maps (SOM) as a visualization tool for materials research and education. Results in Materials, 4, 100020. https://doi.org/10.1016/j.rinma.2019.100020
Potok, T. E., Schuman, C., Young, S., Patton, R., Spedalieri, F., Liu, J., Yao, K. T., Rose, G., & Chakma, G. (2018). A study of complex deep learning networks on high-performance, neuromorphic, and quantum computers. ACM Journal on Emerging Technologies in Computing Systems, 14(2), 1–21.
Osman, A. H., Aljahdali, H. M., Altarrazi, S. M., & Ahmed, A. (2021). SOM-LWL method for identification of COVID-19 on chest X-rays. PLoS ONE, 16(2), e0247176. https://doi.org/10.1371/journal.pone.0247176
Nilashi, M., Ahmadi, H., Sheikhtaheri, A., Naemi, R., Alotaibi, R., Alarood, A. A., Munshi, A., Rashid, T. A., & Zhao, J. (2020). Remote tracking of Parkinson’s disease progression using ensembles of deep belief network and self-organizing map. Expert Systems with Applications, 159, 113562. https://doi.org/10.1016/j.eswa.2020.113562
Khanzadeh, M., Chowdhury, S., Tschopp, M. A., Doude, H. R., Marufuzzaman, M., & Bian, L. (2019). In-situ monitoring of melt pool images for porosity prediction in directed energy deposition processes. IISE Transactions, 51(5), 437–455.
Jafari-Marandi, M., Khanzadeh, W., Tian, B., & Smith, L. (2019). Bian, From in-situ monitoring toward high-throughput process control: Cost-driven decision-making framework for laser-based additive manufacturing. International Journal of Industrial and Manufacturing Systems Engineering, 51, 29–41.
Li, W., Zhang, S., & He, G. (2013). Semisupervised distance-preserving self-organizing map for machine-defect detection and classification. IEEE Transactions on Instrumentation and Measurement, 62(5), 869–879.
Riveros, N. A. M., Espitia, B. A. C., & Pico, L. E. A. (2019). Comparison between K-means and Self-Organizing Maps algorithms used for diagnosis spinal column patients. Informatics in Medicine Unlocked, 16, 100206.
Riese, F. M., Keller, S., & Hinz, S. (2019). Supervised and semi-supervised self-organizing maps for regression and classification focusing on hyperspectral data. Remote Sensing, 12(1), 7.
Mallapragada, S., & Hung, C.-C. (2021). Statistical perspective of SOM and CSOM for hyper-spectral image classification. IEEE Explorer. https://doi.org/10.1109/IGARSS39084.2020.9324200
Shiny, X. A., & Kannan, R. J. (2015). Energy efficient clustering protocol using self organizing map in MANET. Indian Journal of Science and Technology, 8(28), 1.
Abdulkadhim, F. G., Yi, Z., Tang, C., et al. (2021). Design and development of a hybrid (SDN + SOM) approach for enhancing security in VANET. Applied Nanoscience. https://doi.org/10.1007/s13204-021-01908-2
Aftab, F., Zhang, Z., & Ahmad, A. (2017). Self-organization based clustering in MANETs using zone based group mobility. IEEE Access, 5, 27464–27476.
Barletta, V. S., Caivano, D., Nannavecchia, A., & Scalera, M. (2020). Intrusion detection for in-vehicle communication networks: An unsupervised Kohonen SOM approach. Future Internet, 12, 119. https://doi.org/10.3390/fi12070119
Deng, Li. (2014). A tutorial survey of architectures, algorithms and applications for deep learning. APSIPA Transactions on Signal and Information processing, 3(2014), 1–29.
Hinton, G. E. (2009). Deep belief networks. Scholarpedia, 4(5), 5947. https://doi.org/10.4249/Scholarpedia.5947
Bengio, Y. (2009). Learning deep architectures for AI. now Publishers Inc. https://doi.org/10.1561/9781601982957
Movahedi, F., Coyle, J. L., & Sejdic, E. (2018). Deep belief networks for Electroencephalography: A review of recent contributions and future outlooks. IEEE Journal of Biomedical and Health Informatics, 2(3), 642–652. https://doi.org/10.1109/jbhi.2017.2727218
Movahedi, F., Coyle, J. L., & Sejdic, E. (2018). Deep belief networks for electroencephalography: a review of recent contributions and future outlooks. IEEE Journal of Biomedical and Health Informatics, 22(3), 642–652. https://doi.org/10.1109/JBHI.2017.2727218
Liu, L. X. & Xiong, C. (2017). Image classification with deep belief networks and improved gradient descent, In: 2017 IEEE International Conference on Computational Science and Engineering (CSE) and IEEE International Conference on Embedded and Ubiquitous Computing (EUC), pp. 375–380, https://doi.org/10.1109/CSE-EUC.2017.74.
Yang, J., Chang, B., Wang, X., Zhang, Q., Wang, C., Wang, F., & Miao, Wu. (2020). Design and application of deep belief network based on stochastic adaptive particle swarm optimization. Mathematical Problems in Engineering, 2020, 1–10. https://doi.org/10.1155/2020/6590765
Wang, D., & Shang, Yi. (2013). Modeling physiological data with deep belief networks. International Journal of Information and Education Technology (IJIET), 3(5), 505–511. https://doi.org/10.7763/IJIET.2013.V3.326
Zhang, X., Wang, R., Zhang, T., Zha, Y. (2016). Short-term load forecasting based on a improved deep belief network, In: Interntional conference on smart grid and clean energy technologies (ICSGCE), pp. 339–342
Dai, X., Junying Cheng, Y., Gao, S. G., Yang, X., Xiaoqian, X., & Cen, Y. (2020). Deep belief network for feature extraction of urban artificial targets. Mathematical Problems in Engineering, 2020, 1–13. https://doi.org/10.1155/2020/2387823
O’Connor, P., Neil, D., Liu, S.-C., Delbruck, T., & Pfeiffer, M. (2013). Real-time classification and sensor fusion with a spiking deep belief network. Frontiers in Neuroscience. https://doi.org/10.3389/fnins.2013.00178
Abdellaoui, M., & Douik, A. (2020). Human action recognition in video sequences using deep belief networks. Traitement du Signal, 37(1), 37–44. https://doi.org/10.18280/ts.370105
Laqtib, S., Yassini, K. E., & Hasnaoui, M. L. (2019). A deep learning methods for intrusion detection systems based machine learning in MANET. In: International Conference.
Tripathi, K. N., Yadav, A. M., & Sharma, S. C. (2022). Fuzzy and deep belief network based malicious vehicle identification and trust recommendation framework in VANETs. Wireless Personal Communications, 124, 2475–2504.
Naskath, J., Paramasivan, B., et al. (2020). A Study on Modeling Vehicles Mobility with MLC for enhancing vehicle-to-vehicle connectivity in VANET. Journal of Ambient Intelligence and Humanized Computing. https://doi.org/10.1007/s12652-020-02559-x
Danilchenko, K., Azoulay, R., Reches, S., & Haddad, Y. (2022). Deep learning method for delay minimization in MANET. ICT Express, 8(1), 7–10.
Naskath, et al. (2018). Location optimization for road side unit deployment and maximizing communication probability in multilane highway. International Journal of Heavy Vehicle Systems, 25, 369.
Aboelfottoh and Azer, M. A. (2022). Intrusion detection in VANETs and ACVs using deep learning, In: 2022 2nd international mobile, intelligent, and ubiquitous computing conference (MIUCC), pp. 241–245, https://doi.org/10.1109/MIUCC55081.2022.9781691.
Liu, Y., Wang, Y., Lombardi, F., & Han, J. (2018). An energy-efficient stochastic computational deep delief network, Design, Automation and Test. In: Europe conference and exhibition (DATE), pp. 1175–1178.
Alzubaidi, L., Zhang, J., Humaidi, A. J., Al-Dujaili, A., Duan, Ye., Omran Al-Shamma, J., Santamaría, M. A., Fadhel, M.A.-A., & Farhan, L. (2021). Review of deep learning: concepts, CNN architectures, challenges, applications, future directions. Journal of Big Data. https://doi.org/10.1186/s40537-021-00444-8
Kohonen, T., Oja, E., Simula, O., Visa, A., Kangas, J. Engineering applications of the self-organizing map. Proceedings of the IEEE, 84(10), 1358–1384. https://doi.org/10.1109/5.537105
Singh, N. H., & Thongam, K. (2018). Mobile robot navigation using MLP-BP approaches in dynamic environments. Arabian Journal for Science and Engineering, 43(12), 8013–8028. https://doi.org/10.1007/s13369-018-3267-2
Behnisch, M., & Ultsch, A. (2009). Urban data-mining: spatiotemporal exploration of multidimensional data. Building Research & Information, 37(5–6), 520–532. https://doi.org/10.1080/09613210903189343
Chifu, E., & Letia, I. (2008). Unsupervised aspect level sentiment analysis using self-organizing maps conference paper, September 2015. https://doi.org/10.1109/SYNASC.2015.75
Ali Alheeti, K. M., & McDonald-Maier, K. (2018). Intelligent intrusion detection in external communication systems for autonomous vehicles. Systems Science & Control Engineering, 6(1), 48–56. https://doi.org/10.1080/21642583.2018.1440260
Funding
Not Applicable.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
Not Applicable.
Consent for publication
We author of the above titled paper hereby declare that the work included in the above paper is original and is an outcome of the research carried out by the authors indicated in it. Further, we author declare that the work submitted for Wireless Personal Communications an International Journal has not been published already or under consideration for publication in any Journals/Conferences/Symposia/Seminars.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Naskath, J., Sivakamasundari, G. & Begum, A.A.S. A Study on Different Deep Learning Algorithms Used in Deep Neural Nets: MLP SOM and DBN. Wireless Pers Commun 128, 2913–2936 (2023). https://doi.org/10.1007/s11277-022-10079-4
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11277-022-10079-4