Abstract
Restricted Boltzmann Machines (RBMs) and models derived from them have been successfully used as basic building blocks in deep artificial neural networks for automatic features extraction, unsupervised weights initialization, but also as density estimators. Thus, their generative and discriminative capabilities, but also their computational time are instrumental to a wide range of applications. Our main contribution is to look at RBMs from a topological perspective, bringing insights from network science. Firstly, here we show that RBMs and Gaussian RBMs (GRBMs) are bipartite graphs which naturally have a smallworld topology. Secondly, we demonstrate both on synthetic and realworld datasets that by constraining RBMs and GRBMs to a scalefree topology (while still considering local neighborhoods and data distribution), we reduce the number of weights that need to be computed by a few orders of magnitude, at virtually no loss in generative performance. Thirdly, we show that, for a fixed number of weights, our proposed sparse models (which by design have a higher number of hidden neurons) achieve better generative capabilities than standard fully connected RBMs and GRBMs (which by design have a smaller number of hidden neurons), at no additional computational costs.
1 Introduction
Since its conception, deep learning (Bengio 2009) is widely studied and applied, from pure academic research to largescale industrial applications, due to its success in different realworld machine learning problems such as audio recognition (Lee et al. 2009), reinforcement learning (Mnih et al. 2015), transfer learning (Ammar et al. 2013), and activity recognition (Mocanu et al. 2015). Deep learning models are artificial neural networks with multiple layers of hidden neurons, which have connections only among neurons belonging to consecutive layers, but have no connections within the same layers. In general, these models are composed by basic building blocks, such as Restricted Boltzmann Machines (RBMs) (Smolensky 1987). In turn, RBMs have proven to be successfully not just providing good initialization weights in deep architectures (in both supervised and unsupervised learning), but also as standalone models in other types of applications. Examples are density estimation to model human choice (Osogami and Otsuka 2014), collaborative filtering (Salakhutdinov et al. 2007), information retrieval (Gehler et al. 2006), or multiclass classification (Larochelle and Bengio 2008). Thus, an important research direction is to improve the performance of RBMs on any component (e.g. computational time, generative and discriminative capabilities).
The main contribution of this paper is to look at the deep learning basic building blocks, i.e. RBMs and Gaussian RBMs (GRBMs) (Hinton and Salakhutdinov 2006), from a topological perspective, bringing insights from network science, an extension of graph theory which analyzes real world complex networks (Strogatz 2001). Firstly, we study the topological characteristics of RBMs and GRBMs, finding that these exhibit a smallworld topology (Watts and Strogatz 1998). We then hypothesize that by constraining the topology to be also scalefree (Barabasi and Albert 1999) it is possible to reduce the size of ordinary RBMs and GRBMs models, as it has been shown by Del Genio et al. (2011) that scalefree networks are sparse. We introduce a method to make smallworld, scalefree topologies while still considering local neighborhoods and data distribution. We dub the resulting models as compleX Boltzmann Machine (XBM) and Gaussian compleX Boltzmann Machine (GXBM), respectively. An interesting finding is that constraining such XBM and GXBM topologies at their inception leads to intrinsically sparse networks, a considerable advantage to typical stateoftheart methods in which sparsity is enforced as an aftermath, that is during testing (exploitation) phase (Swersky et al. 2012; Wan et al. 2015; Ranzato et al. 2008; Luo et al. 2011; Lee et al. 2008). In turn, XBM and GXBM have a considerably smaller number of weights, which further on contributes to considerably faster computational times (proportional to the number of weights in the model), both in the training and testing phases. What is more, we found that the proposed topology imposes an inductive bias on XBMs and GXBMs, which leads to better statistical performance than RBMs and GRBMs. Our comparative study is based on both simulated and realworld data, including the Geographical origin of music dataset (Zhou et al. 2014), the MNIST digits dataset, CalTech 101 Silhouettes dataset (Marlin et al. 2010), and the 8 datasets from UCI evaluation suite (Larochelle and Murray 2011). We show that, given the same number of hidden neurons, XBM and GXBM have similar or relatively close capabilities to RBM and GRBM, but are considerably faster thanks to their reduced amount of weights. For instance, in a network of 100 visible and 100 hidden neurons, the reduction in weights was by one order of magnitude. A network with 1000 visible and 1000 hidden neurons led to a reduction in weights by two orders of magnitude. Additionally, we show that given the same amount of weights, RBMs or GRBMs derived models with a higher number of hidden neurons and sparse connectivity achieve better generative capabilities than fully connected RBMs or GRBMs with a smaller amount of hidden neurons.
The remaining of this paper is organized as follows. Section 2 presents background knowledge about Boltzmann machines and complex networks for the benefit of the nonspecialist reader and highlights the key motivations of our work. Section 3 discusses the relation between deep leaning and network science and details the mathematical models of our proposed methods. Section 4 describes the experiments performed and analyzes the results. Finally, Sect. 5 concludes the paper and presents directions of future research.
2 Background and motivations
2.1 Boltzmann machines
Originally derived by Ackley et al. (1985), a Boltzmann machine is a network of symmetrically connected stochastic binary units (or neurons). To formalize a Boltzmann machine, and its variants, three main ingredients are required, namely an energy function providing scalar values for a given configuration of the network, the probabilistic inference and the learning rules required for fitting the free parameters. This bidirectional connected network with stochastic nodes has no unit connected with itself. However, Boltzmann machines with unconstrained connectivity are infeasible in practical problems due to the intractable inference. A critical step in taming computational complexity is to add constraints to the connectivity network, which is what makes Boltzmann machines applicable to real world problems.
Smolensky (1987) presented restricted Boltzmann machine that could learn a probability distribution over its set of inputs. The model architecture was restricted by not allowing intralayer connections between the units, as depicted in Fig. 2 (left). Since their conception, different types of Boltzmann machines have been developed and successfully applied. Yet most of these variations preserve some fundamental characteristics. RBMs are generative stochastic neural networks which consists of two binary layers, the visible layer, \({\mathbf {v}}=[v_1, v_2, \ldots , v_{n_v}]\), and the hidden layer, \({\mathbf {h}}=[h_1, h_2, \ldots , h_{n_h}]\), with \(n_v\) being the number of visible neurons and \(n_h\) the number of the hidden ones. Formally, the energy function of RBMs for any state \(\{{\mathbf {v}},{\mathbf {h}}\}\) is computed by summing over all possible interactions between neurons, weights and biases, as follows:
where \(W_{ij}\) denotes the connection between the visible neuron i and the hidden neuron j, \(a_{i}\) is the bias for visible neuron i and \(b_{j}\) is the bias for hidden neuron j. The term \(\sum _{i,j} v_{i}h_{j}W_{ij}\) represents the total energy between neurons from different layers, \(\sum _{i}v_{i}a_{i}\) represents the energy of the visible layer and \(\sum _{j}h_{j}b_{j}\) the energy of the hidden layer. The inference in RBMs is stochastic. For any hidden neuron j the conditional probability is given by \(p(h_j{\mathbf {v}})={\mathcal {S}}(b_{j}+\sum _{i}v_{i}W_{ij})\), and for any visible unit i it is given by \(p(v_i{\mathbf {h}})={\mathcal {S}}(a_{i}+\sum _{j}h_{j}W_{ij})\), where \({\mathcal {S}}(\cdot )\) is a sigmoid function.
Later on, Hinton and Salakhutdinov (2006) have extended the RBMs models to make them suitable for a large number of applications with realvalued feature vectors. They used exponential family harmoniums results from Welling et al. (2005) and developed the Gaussian Restricted Boltzmann Machine (GRBM) model that, like RBMs, forms a symmetrical bipartite graph. However, the binary units from the visible layer \({\mathbf {v}}\) are replaced by linear units with Gaussian noise. The hidden units \({\mathbf {h}}\) remain binary. Therein, the total energy function for a state \(\{{\mathbf {v}},{\mathbf {h}}\}\) of GRBMs is calculated in a similar manner to RBMs, but includes a slight change to take into consideration the Gaussian noise of the visible neurons, as defined in Eq. 2.
where, the term \(\sum _{i,j} \frac{v_{i}}{\sigma _{i}}h_{j}W_{ij}\) gives the total energy between neurons from different layers, \(\sum _{i}\frac{(v_{i}a_{i})^2}{2\sigma _{i}^2}\) is the energy of the visible layer, and \(\sigma _i\) represents the standard deviation of the visible neuron i. The stochastic inference for any hidden neuron j can be done as for RBMs, while for any visible unit i is made by sampling from a Gaussian distribution, defined as \({\mathcal {N}}(a_{i}+\sum _{j}h_{j}W_{ij},\sigma _{i}^2)\).
Parameters of RBM and GRBM models are fitted by maximizing the likelihood function. In order to maximize the likelihood of the model, the gradients of the energy function with respect to the weights have to be calculated. Because of the difficulty in computing the derivative of the loglikelihood gradients, Hinton (2002) proposed an approximation method called Contrastive Divergence (CD). In maximum likelihood, the learning phase actually minimizes the Kullback–Leiber (KL) measure between the input data distribution and the model approximation. Thus, in CD, learning follows the gradient of:
where, \(p_{n}(.)\) is the resulting distribution of a Markov chain running for n steps. Besides that, other methods have been proposed to train RBMs (e.g. persistent contrastive divergence (Tieleman 2008), fast persistent contrastive divergence (Tieleman and Hinton 2009), parallel tempering (Desjardins et al. 2010)), or to replace the Gibbs sampling with a transition operator for a faster mixing rate and to improve the learning accuracy without affecting computational costs (Brgge et al. 2013).
2.2 Sparsity in restricted Boltzmann machines
In general and for the purposes of machine learning, obtaining a sparse version of a given model, leads to a reduction in parameters, which, in turns helps in addressing problems such as overfitting and excessive computational complexity. The sparsity issue in RBMs is so important that considerable attention is given to it in the literature. Hereafter we point to the most relevant works but will not attempt to provide a comprehensive overview. It is worth mentioning the work by Dieleman and Schrauwen (2012) and Yosinski and Lipson (2012) who have shown how the histogram of the RBM weights changes shape during the training process, going from a Gaussian shape (initially) to a shape that peaks around zero (which provides a further motivation towards sparsity enforcement).
One of the most common methods to obtain sparse representations is by encouraging it during the training phase using different variants of a sparsity penalty function, as done for instance, in Lee et al. (2008) and Ranzato et al. (2008). However, performing this process during the learning phase does not guarantee sparsity in the testing phase (Swersky et al. 2012). To overcome this limitation, CardinalityRBM (CaRBM) was proposed by Swersky et al. (2012) to ensure sparsity in the hidden representation by introducing cardinality potentials into the RBMs energy function. Moreover, Wan et al. (2015) have proposed Gaussian CaRBM, in which they replace the universal threshold for hidden units activation from CaRBM with adaptable thresholds. These thresholds are sampled from a certain distribution which takes into consideration the input data. Recently, Han et al. (2015) introduced one of the most efficient methods to obtain weights sparsity in deep neural network. They successfully obtained up to 10 times less weights in deep neural networks with no loss in accuracy. The method assumes three simple steps: (1) the network is trained to learn the most important connections; (2) the unimportant connections are pruned; (3) the network is retrained to fine tune the weights of the remaining connections. To achieve the best performance the steps 2 and 3 have to be repeated iteratively making it a computationally expensive method.
Thus, to the best of our knowledge, all of the stateoftheart methods impose a sparsity regularization target during the learning process, which makes them impractical with large datasets having millions (or even billions) of input features. This is because the training process is excessively slow in such situations. Our solution to overcome this problem is to ensure weight sparsity from the initial design of an RBM using relevant findings from the field of network science. To this end, next section introduces these findings.
2.3 Complex networks
Complex networks (e.g. biological neural networks, actors and movies, power grids, transportation networks) are everywhere, in different forms and different fields, from neurobiology to statistical physics (Strogatz 2001), and they are studied in network science. Formally, a complex network is a graph with non trivial topological features, human or nature made. The most two wellknown and deeply studied types of topological features in complex networks are the scalefree and the smallworld concepts, due to the fact that a wide range of realworld complex networks have these topologies. A network with a scalefree topology (Barabasi and Albert 1999) is a sparse graph (Del Genio et al. 2011) that approximately has a powerlaw degree distribution \(P(k)\sim k^{\gamma }\), where the fraction P(k) from the total nodes of the network has k connections to other nodes, and the parameter \(\gamma \in (2,3)\) usually.
At the same time, a network model with the smallworld topological feature (Watts and Strogatz 1998) is defined to be a graph in which the typical distance (L) between two randomly chosen nodes (the number of hops required to reach one node from the other) is very small, approximately on the logarithmic scale with respect to the total number of nodes (N) in the network, while at the same time it is characterized by a clustering coefficient which is significantly higher than may appear by random chance. More formally, a graph sequence \((G_N)_{N\ge 1}\) has a smallworld topology, if there is a constant \(0<K<\infty \) such that \(lim_{N\rightarrow \infty }p(L_N\le K log N)=1\), where \(L_N\) is the typical shortest path of \(G_N\) (van der Hofstad 2016). As an example, Fig. 1 roughly illustrates a smallworld topology, and a scalefree one, in two small randomly generated graphs. Both types of topologies are studied below in the context of restricted Boltzmann machines, leading to our proposal of sparse Boltzmann machine models.
3 Complex networks and Boltzmann machines
In this section, we firstly discuss the relation between complex networks on one side and restricted Boltzmann machines and Gaussian restricted Boltzmann machine on the other side. Secondly, we introduce an algorithm to generate sparse topologies for bipartite graphs which have both properties (i.e. scalefree and smallworld) that also considers the distribution of the training data. Finally, we make use of the previous mentioned topology generator algorithm and present the mathematical details of two novel types of Boltzmann machines, dubbed compleX Boltzmann Machines (XBMs) and Gaussian compleX Boltzmann Machines (GXBMs).
3.1 Topological insight into RBMs and GRBMs
Lately, the neural networks of the human brain have started to be studied using tools from network science (Pessoa 2014). It has been found that these exhibit both a smallworld topology (i.e. the shortest path between any two nodes or neurons is very small, approximately equal to the logarithm of the total number of nodes) and a scalefree topology (i.e their degree distribution follows a power law). At the same time, by making small steps towards mimicking the architecture and the functionality of the brain, deep learning methods have emerged as a promising solution in computer science to develop automated learning systems (Jones 2014; Mnih et al. 2015). Here we argue that there is a clear relation between network science and deep learning. In the scope of these arguments, we introduce the following proposition:
Proposition 1
Both restricted Boltzmann machines and Gaussian restricted Boltzmann machines are bipartite graphs which have a smallworld topology.
Proof
The diameter (i.e. the longest shortest path between any two neurons) of RBMs or GRBMs is 2, independently on the number of hidden or visible neurons, due to the fact that both models have all the possible interlayer connections, but no intralayer connections. This yields that L is bounded up by 2 for any RBM or GRBM. By replacing L in the smallworld definition from Sect. 2.3, we obtain \(lim_{N\rightarrow \infty }p(2\le K log N)=1\), which is true for any constant K, \(0<K<\infty \).^{Footnote 1} Similarly as RBMs and GRBMs are complete bipartite graphs, their clustering coefficient (Latapy et al. 2008) is 1, being higher than any other possible cluster coefficient.^{Footnote 2} Thus, it is clear that any RBMs or GRBMs have a smallworld topology. \(\square \)
Following the same line, our intuition is that by preserving the smallworld property of RBMs or GRBMs, while introducing also the scalefree property in their topology, we can obtain new sparse Boltzmann machines derived models which may have similar performance to RBMs or GRBMs, but with fewer free parameters (i.e. the weights between the visible and the hidden neurons). Thus, further on, we introduce the complex Boltzmann machine and the Gaussian complex Boltzmann machine (i.e. the derivatives of RBM and GRBM, respectively) which exhibit both scalefree and smallworld topological properties.
3.2 Topology generation algorithm for XBM and GXBM
To generate a sparse topology in XBM and GXBM, we have devised a three stages heuristic method, detailed in Algorithm 1. In the first stage, a scalefree bipartite graph is generated; in the second one, the graph is adjusted to be also smallworld; and in the third stage, the graph topology is fitted to the data distribution. Below, this method is thoroughly discussed.
First we generate a powerlaw degree sequence with \(n_v+n_h\) elements, using \(P(k)=k^{\gamma }\), \(\forall k \in {\mathbb {N}}\), \(1\le k \le n_v+n_h\) , with minimum degree equal to four to favor the smallworld topology, and we sort it in descending order (i.e. Algorithm 1, lines 8–9). Each element from the sequence represents a node in the network and the actual value of that element (i.e. the degree) represents the number of connections for that specific node. After that, we split the degree sequence in two, by alternatively picking the nodes starting from those with the highest degree and proceeding until the smallest degree nodes are reached. In this way we populate two separate lists, for the hidden layer and for the visible layer, respectively (i.e. Algorithm 1, lines 10–16). Once the list having the smallest number of elements is completed, we add all the remaining elements of the original sequence to the bigger list (i.e. Algorithm 1, lines 17–22). In an undirected bipartite graph both layers need to have an equal number of connections. Due to the fact that the sum of the elements from one list might not be equal with the sum of the elements from the other list, we add some more degrees to the elements of the list with less degrees (proportionally to its initial degree distribution) to equalize the two lists (i.e. Algorithm 1, lines 23–28). Next, starting from these two lists, we create a bipartite graph G using a HavelHakimi procedure (Hakimi 1962) (i.e. Algorithm 1, line 29). Further on, we add few more connections to each node in G with the aim to achieve the optimal clustering coefficient as required in smallworld topologies (i.e. Algorithm 1, lines 30–49). This ensures also a dense local connectivity useful, by example, for images. To clarify how this is done, note that in Algorithm 1 (line 3^{Footnote 3}), when parameter \(\sigma ^{neigh}\) is increased the nodes local neighborhoods gradually turn into larger neighborhoods (or even the graph as a whole). In turn, when parameter \(\phi \) is increased, the neighborhoods tend to become denser. The whole algorithm proceeds iteratively until the bipartite graph meets the criteria of smallworldness (i.e. Algorithm 1, line 51). We should mention, though, that during our whole study we observed that this property was usually achieved after just one iteration. To take the data distribution into consideration, as a final stage, the algorithm rearranges the order of the visible nodes in G such that the nodes having more connections end up corresponding to the training data features with a higher standard deviation (i.e. Algorithm 1, line 53). The resulting bipartite graph G can then be used as the topology of our XBM or GXBM models (i.e. Algorithm 1, line 55–57), as detailed next.
3.3 Complex Boltzmann machines
Just like restricted Boltzmann machines, complex Boltzmann machines are made up, by two layers of neurons, with connections only in between different layers and no connections within the same layer. The bottom layer (i.e. the visible one) is denoted further with a binary vector \({\mathbf {v}}=[v_1, v_2, \ldots , v_{n_v}]\), in which each unit \(v_i\) is a binary unit, and where \(n_v\) is the size of \({\mathbf {v}}\) (i.e. the number of neurons of the visible layer). The top layer (i.e. the hidden one) is represented further by the binary vector \({\mathbf {h}}=[h_1, h_2, \ldots , h_{n_h}]\), in which each element \(h_j\) is binary, and where \(n_h\) is the size of \({\mathbf {h}}\). Furthermore, each neuron from the visible layer has associated one bias. The biases of the visible neurons are collected in a vector \({\mathbf {a}}=[a_1, a_2, \ldots , a_{n_v}]\). Similarly, hidden layer neurons have biases, collected in vector \({\mathbf {b}}=[b_1, b_2, \ldots , b_{n_h}]\).
The difference between RBM and XBM consists in how the neurons from the different layers are connected between them. RBMs form a full undirected mesh between all the neurons on the hidden layer and all the neurons on the visible layer. By contrast, in XBMs the connections between the two layers are still undirected but sparse, as generated by Algorithm 1. Thus, XBMs have both scalefree and smallworld topological properties. These connections are defined in a sparse adjacency weights matrix \({\mathbf {W}}=[[w_{11} ,w_{12}, \ldots , \) \({w_{1n_h}}], \ldots , [w_{{n_{v}1}}, w_{{n_{v}2}}, \ldots , {w_{{n_{v}n_{h}}}}]]\) in which the elements are either null (\(w_{ij}=0\)) when there is no connection between the visible neuron i and the hidden neuron j or have a connection weight (\(w_{ij}\ne 0\)) when the connection between i and j exists. The high level architectures of RBMs and XBMs are depicted in Fig. 2. The sparse topology of XBMs leads to a much smaller number of connections, which further on leads to faster computational times than RBMs.
3.3.1 The XBM energy function
The energy function of an XBM is defined as:
where, \(\mathbf \Gamma _i^h\) is the set of all hidden neurons connected to the visible neuron i (i.e. \(\mathbf \Gamma _i^h=\{j1\le j \le n_h, \forall j \in {\mathbb {N}} \wedge w_{ij}\ne 0\)}).
3.3.2 XBM inference
Due to the fact that there are no links between the neurons on the same layer, inference can still be performed in parallel for any visible neuron i or any hidden neuron j, as below:
where, \(\mathbf \Gamma _j^v\) is the set of all visible neurons connected to the hidden neuron j (i.e. \(\mathbf \Gamma _j^v=\{i1\le i \le n_v, \forall i \in {\mathbb {N}} \wedge w_{ij}\ne 0\)}), \({\mathcal {S}}(\cdot )\) is the sigmoid function, and \(\mathbf {\Theta }\) represents the free parameters of the model (i.e. \({\mathbf {W}}\), \({\mathbf {a}}\), \({\mathbf {b}}\)).
3.3.3 XBM learning
The general update rule for the free parameters \(\mathbf {\Theta }\) of the GXBM model is given by:
where \(\tau \), \(\rho \), \(\alpha \), and \(\xi \) represent the update number, momentum, learning rate, and weights decay, respectively. For a thorough discussion on the optimal choice of these parameters the interested reader is referred to Hinton (2012). Furthermore, \(\nabla \mathbf {\Theta }_{\tau +1}\) for each of the free parameters can be computed by using contrastive divergence (Hinton 2002) and deriving the energy function from Eq. 4 with respect to that parameter, yielding:
with \(\langle \cdot \rangle _n\) being the distribution of the model obtained after n steps of Gibbs sampling in a Markov Chain which starts from the original data distribution \(\langle \cdot \rangle _0\). We must note that in this paper we have chosen to train our proposed models using the original contrastive divergence method (Hinton 2002), which is widely used and allows for a direct comparison to the results reported in the literature. It may well be that other training methods would offer better performance; yet, overall, we do not expect that a particular training method to significantly affect our findings.
3.4 Gaussian complex Boltzmann machines
Just like in GRBMs and RBMs, the only differences between GXBMs and XBMs is that in the case of GXBMs the visible layer \({\mathbf {v}}=[v_1, v_2, \ldots , v_{n_v}]\) has real values and each \(v_i\) is a linear unit with Gaussian noise (Hinton and Salakhutdinov 2006). Thus, the total energy equation of GXBMs is slightly changed to reflect the real visible layer, as follows:
where, \(\sigma _i\) represents the standard deviation of the visible neuron i. We note that in the remainder we adopt the same notations used in XBM modeling, unless specified otherwise.
Furthermore, the inference in GXBMs can still be performed in parallel for any visible neuron i or any hidden neuron j, as below:
where, \({\mathcal {N}}(\cdot ,\cdot )\) represents a Gaussian distribution. Finally, the learning in GXBM can be done using the same procedure as for XBM (Sect. 3.3.3).
4 Experimental results
To assess the performance of XBM and GXBM we have conducted three sets of experiments in a stepwise fashion. In the first one, we study the behavior of XBM, GXBM and their topology generation algorithm. In the second one, we analyze the reconstruction error obtained by GXBM on random generated data and on a real world dataset, more exactly on the Geographical Origin of Music dataset (Zhou et al. 2014). Thirdly, we assess the statistical performance of XBM on the MNIST digits dataset, CalTech 101 Silhouettes dataset (Marlin et al. 2010), and the 8 datasets from UCI evaluation suite (Larochelle and Murray 2011) using Annealed Importance Sampling (AIS) (Salakhutdinov and Murray 2008).
Furthermore, in the last two sets of experiments, we compare GXBM/XBM against three other methods, as follows: (1) the standard fully connected GRBM/RBM; (2) sparse GRBM/RBM models, denoted further GRBM\(_\text {FixProb (Fixed Probability)}\)/RBM\(_\text {FixProb}\), in which the probability for any possible connection to exist is set to the number of weights of the counterpart GXBM/XBM model divided by the total number of possible connection for that specific configuration of hidden and visible neurons;^{Footnote 4} and (3) sparse GRBM/RBM models, denoted further GRBM\(_\text {TrPrTr (Train Prune Train)}\)/RBM\(_\text {TrPrTr}\), in which the sparsity is obtained using the algorithm introduced in Han et al. (2015) with L2 regularization, and in which the weights sparsity target is set to the number of weights of the counterpart GXBM/XBM model. Please note that in all experiments if the weights sparsity target was not reached after 50 pruning iterations, we stopped the training algorithm and we used the obtained GRBM\(_\text {TrPrTr}\)/RBM\(_\text {TrPrTr}\) model. For each evaluated case, we analyze two scenarios when: (1) the number of connections is the same for the sparse and the full connected models, while the number of hidden neurons is different; (2) the number of hidden neurons is the same for the sparse and the full connected models, while the number of connections is different.
4.1 Scrutinizing XBM and GXBM topologies
In this set of experiments, we analyze the characteristics of the XBM sparse topology. For the sake of brevity, we refer just to the XBM–RBM relation, since the GXBM–GRBM is identical from the topological point of view. To perform the various operations on the bipartite graph we have used the NetworkX library (Hagberg et al. 2008), setting the power law coefficient \(\gamma \) to 2, a typical value in various realworld networks (Clauset et al. 2009). Firstly, we verified the output of the topologygeneration algorithm (i.e. Algorithm 1). Being one of the algorithm constraints, the smallworld property is preserved in XBM. At the same time, due to the fact that the neurons from different layers are connected starting from a power law degree sequence, the scalefree property is also preserved. As example, Fig. 3a depicts the weights distribution for an XBM having 784 visible and 1000 hidden neurons, while Fig. 3b shows the degree distribution of an XBM with 100 visible and 1000 hidden neurons on the loglog scale. Figure 3b exhibits evidently a scalefree degree distribution. All other experiments exhibited the required scalefree distribution. Furthermore, we analyzed the number of connections in XBM in comparison with the number of connections in RBMs, given the same number of hidden (\(n_h\)) and visible (\(n_v\)) neurons. Figure 3c depicts how many times the number of connections in RBM is bigger than the number of connections in XBM for various configurations (the number of hidden and visible neurons varies from 10 to 1000). The actual values in the heat map are computed using the following formula \(n_w^{{\textit{RBM}}}/n_w^{{\textit{XBM}}}\), where \(n_w^{{\textit{XBM}}}\) is obtained after counting the links given by the topology generation algorithm for XBM, and \(n_w^{{\textit{RBM}}}=n_vn_h\). It can be observed that as the number of hidden and visible neurons increases, the number of weights in XBM becomes smaller and smaller than the one in RBM. For instance, we achieve around 14 times less weights in XBM for 100 visible and 100 hidden neurons, and approximatively 95 times less weights in XBM for 1000 visible and 1000 hidden neurons.
4.2 GXBM evaluation
In the second set of experiments, we assess the performance of GXBM against GRBM, GRBM\(_\text {FixProb}\), and GRBM\(_\text {TrPrTr}\) on randomly generated as well as on the realworld dataset Geographical Origin of Music (Zhou et al. 2014).
Settings and implementations For all experiments performed in this set we have used Python implementations of the four models under scrutiny. In all models, the momentum was set to 0.5, the learning rate to 0.001, and the number of Gibbs sampling in contrastive divergence to 1, as discussed by Hinton (2012). The weights decay was 0.0002 for GXBM, GRBM, and GRBM\(_\text {FixProb}\), while for GRBM\(_\text {TrPrTr}\) we used the L2 regularization. The number of neurons in the visible layer was set to the dimension of the input data, while the number of hidden neurons was varied for a better comparison. In the learning phase, we stopped the models after 100 training epochs to ensure full convergence. In fact, convergence was much faster, as exemplified in Fig. 3d which shows the case of GXBM with 100 visible and 1000 hidden neurons trained on random generated data. In the case of GRBM\(_\text {TrPrTr}\), we have repeated the training procedure for a maximum of 50 pruning iterations trying to reach the same amount of weights as in GXBM, but this target was impossible to reach in all situations.
Performance metrics To quantify the performance of the models, we used a variety of standard metrics. We have used: the Root Mean Square Error (RMSE) to estimate the distance between the reconstructed inputs and the ground truth; the Pearson Correlation Coefficient (PCC) to reflect the correlations between the estimated inputs and the ground truth; and the Pvalue to arrive at a statistically significant reconstruction level during the learning phase.
4.2.1 GXBM performance on random generated data
Firstly, we analyze how well GXBM is capable to reconstruct random generated data. To this end, we have generated 1000 data points, each one having 100 dimensions, and each dimension being sampled from a Gaussian distribution with 0 mean and standard deviation equal to 1, i.e. \({\mathcal {N}}(0,1)\). Due to the fact that these are random generated data, there was no reason to use cross validation. Thus, we have used \(70\,\%\) of data to train the models, and the remaining \(30\,\%\) to test the models. Firstly, we analyzed the reconstruction capabilities of GXBM, GRBM\(_\text {FixProb}\), GRBM\(_\text {TrPrTr}\), and GRBM, given the the same number of weights. Figure 4 (left) depicts this situation, while the number of weights were varied from 700 up to approximately 7000. Clearly, GXBM outperforms GRBM in both, RMSE and PCC, while its internal topology permits it to have a higher number of hidden neurons. Remarkably, with approximately 1000 weights, the mean RMSE for GXBM is already very low, around 0.3, while the mean PCC is almost perfect, over 0.95. By contrast, GRBM with 1000 weights performed poorly. Furthermore, it is clear that the GRBM performance increases with the number of weights, yet it can be observed that even at approximately 7000 weights GRBM is not capable to reach the same level of performance of the GXBM with 1000 weights. Besides that, GXBM outperforms also the other sparse models, GRBM\(_\text {FixProb}\) and GRBM\(_\text {TrPrTr}\), but not so drastically. In fact, GRBM\(_\text {FixProb}\) has very close performance to GXBM. GRBM\(_\text {TrPrTr}\) is not so close, while having also a very high computational cost as depicted in the fourth row of Fig. 4.
To better understand these differences, we proceed to the next scenario, analyzing GRBM, GRBM\(_\text {FixProb}\), GRBM\(_\text {TrPrTr}\), and GXBM having the same number of hidden neurons. This is reflected in Fig. 4 (right) in which the number of hidden neurons is varied from 100 to 1000. Surprising, even though the number of free parameters (i.e. weights) was smaller by at least one order of magnitude in GXBM (as it can be seen in the bottomright plot of Fig. 4), GXBM performs similarly to GRBM in terms of PCC and RMSE, while GRBM\(_\text {FixProb}\) and GRBM\(_\text {TrPrTr}\) reach almost a similar performance. Still, GRBM\(_\text {TrPrTr}\) has the downside of not being capable to reach the same number of weights as GXBM or GRBM\(_\text {FiXProb}\) even after 50 prunning iterations, as reflected on the fourth and fifth rows of Fig. 4. Interestingly, for this specific dataset, all models seem to reach their maximum learning capacity when they have approximately 400 hidden neurons, showing no further improvement after this point.
4.2.2 GXBM performance on geographical ethnomusicology data
We have then assessed the reconstruction capabilities of GXBM on a real world dataset. We have used the Geographical Origin of Music dataset (Zhou et al. 2014). This contains 1059 tracks from different countries, each track having 70 dimensions (i.e. the first 68 represent audio features, while the last ones are latitude and longitude of each specific song), already normalized to have mean 0 and standard deviation equal to 1. We performed a 10fold cross validation. The averaged results depicted in Fig. 5 confirm the same findings obtained on the random generated dataset (Sect. 4.2.1). Given the same number of weights, GXBM outperforms clearly GRBM, and outperforms slightly GRBM\(_\text {FixProb}\) and GRBM\(_\text {TrPrTr}\), while the four perform similarly when the number of hidden neurons is comparable. By analyzing the reconstruction performance metrics (i.e. RMSE and PCC) reported in Figs. 5 and 4 it is interesting to observe that GXBM and GRBM\(_\text {FixProb}\) are more stable, independently if the data are random or nonrandom. By contrast, GRBM and GRBM\(_\text {TrPrTr}\) performance depend more on the data type.
To assess GXBM from a different perspective, we performed a small regression experiment on the geographical ethnomusicology dataset, even though the regression task does not constitute one of the goals of this paper. Thus, we compared the ability of GRBM and GXBM in predicting the latitude and the longitude corresponding to the 68 audio features across all the tracks. We can see from Fig. 5 that the dataset is well represented with just about 400 hidden neurons; thus we have used this value in both models. In addition, in GXBM we have set the visible neurons corresponding to latitude and longitude to the first two visible neurons having the largest number of connections. We then performed a 10fold cross validation to obtain the average distance error between the predicted latitude and longitude and their true value counterparts. The resulting predictions were \(3258\pm 175\) km (GRBM) and \(3252\pm 176\) km (GXBM) which are comparable to the top performers found in Zhou et al. (2014), even if it is well known that GRBMs and RBMs are not best performers on classification and regression tasks, when used as standalone models—best performance would pursued by stacking these models in deep architectures (Bengio 2009).
From the third row of Figs. 4 and 5, it is very interesting to observe that, even if it was not a target, the random generated connections of GRBM\(_\text {FixProb}\) exhibit a very small average shortest path. This observation may explain the good performance of GRBM\(_\text {FixProb}\) close to the one of GXBM in this set of experiments, while having a similar computational time (i.e. the same number of weights to be computed). Even more, the GRBM\(_\text {TrPrTr}\) models end up still having a smallworld topology after the iterative pruning process. Still, the better performance obtained by GXBM is given by, alltogether, its scalefree topology, the local neighborhoods connections, and the consideration of data distribution in the topology.
4.3 XBM evaluation
In the third set of experiments, we have assessed the performance of XBM on the MNIST digits dataset,^{Footnote 5} CalTech 101 Silhouettes dataset (Marlin et al. 2010), and the 8 datasets from the UCI evaluation suite (Larochelle and Murray 2011).
Settings, implementations, and performance metrics To allow for a direct comparison, we have adopted the same settings for all datasets. This time, all the models were implemented and the experiments were performed using the MATLAB\(^\circledR \) environment, partially to facilitate the comparisons with the results reported on the MNIST dataset in Salakhutdinov and Murray (2008). For each RBM model, we have trained two XBM, two RBM\(_\text {FixProb}\), and two RBM\(_\text {TrPrTr}\) models, as follows: (1) one having the same number of hidden neurons, but with much fewer weights; (2) the other with approximatively the same number of weights but with a higher number of hidden neurons. Just as in Salakhutdinov and Murray (2008), we have used a fixed learning rate (i.e. 0.05) for all datasets, with the exception of the MNIST dataset when a decreasing learning rate was used for the situation in which the number of contrastive divergence steps was gradually increased from 1 up to 25, as suggested in CarreiraPerpinan and Hinton (2005). In all cases, the number of training epochs was set to 259 and the training and testing data were split in minibatches of 100 samples. Finally, to assess the performance of XBM, RBM\(_\text {FixProb}\), and RBM\(_\text {TrPrTr}\), we have used AIS (adopting the very same settings as in Salakhutdinov and Murray 2008) to estimate the average logprobabilities on the training and testing data of the datasets.
4.3.1 XBM performance on the MNIST digits dataset
The MNIST digits dataset is widely used to assess the performance of novel machine learning algorithms. The dataset contains 60000 images for training and 10000 images for testing, each image being a variant of a digit represented by a 28x28 binary matrix.
The results depicted in Table 1 ^{Footnote 6} confirm the findings from the previous set of experiments (i.e. GXBM–GRBM case). We see that: (1) given the same number of hidden neurons an XBM model has a generative performance close to an RBM, yet having the significant benefit of much fewer weights (i.e. this offers much smaller computational time); (2) given approximatively the same number of weights (i.e. at comparable computational time) an XBM model has a better generative performance than an RBM model. The results suggest that XBM models can achieve good generative performance already with one step contrastive divergence and a fix learning rate. This is because the performance gain obtained by increasing the number of contrastive divergence steps is smaller than in the case of RBMs. This may be considered also an advantage for XBMs as it is a common practice to use just one step contrastive divergence to decrease the learning time. It is worth highlighting, that an XBM with 15702 weights reaches 36.02 nats better than an RBM with 15680 weights and 20 hidden neurons. Similarly to the GXBM behaviour, as the models increase in size, we observe that the difference between XBMs and RBMs gets smaller. For instance, an XBM with 387, 955 weights is 10.89 nats better than an RBM with 392, 000 weights, which, further on, is just 6.99 nats better than an XBM having the same number of hidden neurons (i.e. 500) but with approximately 40 times fewer weights (i.e. 10, 790). Also worth noting that when the CD learning steps were gradually increased from 1 to 25 the XBM model with 387, 955 weights slightly outperformed (i.e. 1.13 nats) the RBM model having 500 hidden neurons and 392, 000 weights trained in Salakhutdinov and Murray (2008). We should note that the latter is considered to be one of the best generative RBM models reported in the literature, to the best of our knowledge.
Finally, we would like to highlight that in 5 out of the 6 cases considered, XBMs outperform all the other models including the sparse ones, while in the remaining case, the best performer is not the fully connected RBM as expected, but still a sparse model,i.e. RBM\(_\text {TrPrTr}\). Yet, the latter one, same as in Sect. 4.2, shows some robustness issues as it performs very badly for a large number of hidden neurons (i.e. 27,000), while its computational time is much higher than for all the other models considered. Remarkably, RBM\(_\text {FixProb}\) obtains very good results in all cases, being behind XBM just from few up to a maximum of approximately 25 nats in the worst case.
4.3.2 XBM performance on the CalTech 101 Silhouettes dataset
To confirm the previous results on a different (more complicated) dataset, further on we assess XBMs generative performance on the CalTech 101 Silhouettes dataset (Marlin et al. 2010). This dataset contains silhouettes of objects extracted from the CalTech 101 image dataset. In total it has 101 classes and two datasets. One with binary images of 28x28 pixels split in a training set of 4100 samples and a testing set of 2307 samples, and one with binary images of 16x16 pixels split in a training set of 4082 samples and a testing set of 2302 samples. As our goal was not to fine tune the four models, but to have a clear direct comparison between them, for each dataset we have used 1 CD step and we considered two evaluation cases, a small RBM (i.e. 25 hidden neurons) and a large one (i.e. 500 hidden neurons). Table 2 confirms our previous findings and shows that XBMs are still the best performers outperforming clearly all the other models, with a striking difference of 118.48 nats against fully connected RBMs (which are subject to overfitting) on the dataset of 28x28 image size. On both datasets, it is interesting to see that the best XBMs performers are not the largest models (i.e. 27,000 and 10,000 hidden neurons, respectively), but the average size ones (i.e. 1500 and 500 hidden neurons, respectively).
4.3.3 XBM performance on the UCI evaluation suite
Up to now all the datasets used to evaluate the XBMs were binarized images. In this last subset of experiments, we use the UCI evaluation suite, which contains 8 binary datasets coming from various domains. These datasets are carefully selected by Larochelle and Murray (2011) to assess the performance of generative and density estimation models. Their characteristics are well described in Germain et al. (2015) and are summarized in Table 3. As before, to have a clear direct comparisons of all sparse models we have compared them with baseline fully connected RBMs with 100 hidden neurons using standard 1 CD step. Table 4 summarizes the results. XBMs outperform all the other models, including the fully connected RBMs, on 7 out of the 8 datasets. As usual, RBM\(_\text {FixProb}\) shows a good performance overall, being even the best performer on one dataset, i.e. Connect4. By contrast, RBM\(_\text {TrPrTr}\) has robustness issues, sometimes showing good generative capabilities and sometimes not. Even if it was not in our goal to outperform the best results from the literature and, as a consequence, we did not fine tune any parameter and we did not try other training algorithms, XBMs reach on all datasets very good performances close to the ones of the best generative models carefully optimized in Germain et al. (2015).
To summarize this set of experiments performed on 10 binary datasets (i.e. MNIST digits, CalTech 101 Silhouettes, and UCI evaluation suite), we report that XBMs outperform all the other models in 16 out of 18 cases considered. In the other two cases, the winner is once RBM\(_\text {FixProb}\), and once RBM\(_\text {TrPrTr}\). Besides that, a very interesting finding is that in all cases RBM\(_\text {TrPrTr}\) models end up having a smallworld topology, as reflected by their average shortest path and cluster coefficient. Similarly, RBM\(_\text {FixProb}\) models reach a very small average shortest path (suitable to be qualified as smallworlds), but we can not consider them pure smallworlds as their average cluster coefficient represents the one obtained by random chance. Still, the better overall performance obtained by XBMs may be explain by the fact that its smallworld topology is supplemented by its other designed topological features, i.e. scalefree property, the consideration of local neighborhoods and data distribution. As reflected by experiments (including the ones with realvalued data for GXBMs) these features complements each other, while helping XBMs to model well very different data types.
5 Conclusion
In this paper we look at the deep learning basic building blocks from a topological perspective, bringing insights from network science. Firstly, we point out that RBMs and GRBMs are smallworld bipartite networks. Secondly, by introducing scalefree constraints in RBMs and GRBMs, while still considering some local neighborhoods of visible neurons, and fitting the most connected visible neurons to the most important data features, we propose two novel types of Boltzmann machine models, dubbed complex Boltzmann machine and Gaussian complex Boltzmann machine. Looking at both artificial and realworld datasets (i.e. Geographical Origin of Music, MNIST digits, CalTech 101 Silhouettes, and UCI evaluation suite) we show that XBM and GXBM obtain better performance than other two sparse models (i.e. RBM\(_\text {FixProb}\)/GRBM\(_\text {FixProb}\) and RBM\(_\text {TrPrTr}\)/GRBM\(_\text {TrPrTr}\)) and we illustrate how they outperform even the fully connected RBM and GRBM, respectively:

1.
Given the same number of hidden neurons, our proposed models exhibit much faster computational time thanks to a smaller number of parameters which have to be computed (up to a few orders of magnitude smaller than in RBM and GRBM) and comparable reconstruction capabilities.

2.
Given the same number of weights, or implicitly a much higher number of hidden neurons for XBM and GXBM, they significantly outperform RBM and GRBM, respectively.
It is worth noting that as the number of neurons increases the order of magnitude between the number of weights in XBM/GXBM and in RBM/GRBM also increases (e.g. one order of magnitude fewer weights in a XBM/GXBM with 100 visible and 100 hidden neurons, and two orders reduction in a XBM/GXBM with 1000 visible and 1000 hidden neurons). This relation will help increasing the typical number of neurons in deep artificial neural networks from the few hundred thousands of today (Krizhevsky et al. 2012; Ba and Caruana 2014) to even billions in the nearfuture. In turn this will lead to the ability to tackle problems having much higher dimensional data—something that is today unfeasible without performing dimensionality reduction. For instance, when working on ordinary images^{Footnote 7} today is still a common practice to first extract features using standard image processing techniques, and just those features can be served as inputs to deep models—an example can be found in Srivastava and Salakhutdinov (2012). We speculate that another significant benefit of using sparse topologies, as in XBM/GXBM, would be the ability to better disentangle the features extracted automatically by the hidden layer.
To conclude, in this article, we have shown empirically on 12 datasets that our proposed models, i.e. XBMs and GXBMs, achieve a very good performance as generative models. We mention that more research has to be done in order to understand why their proposed topology (e.g. the scalefree constraints) makes them to perform so well. Further on, we intend to investigate all these directions and to study analytically how the various parameters of the topologygeneration algorithm (implicitly the bipartite graph properties) may increase or decrease the generative and discriminative capabilities of XBMs and GXBMs, respectively.
Notes
Please note that according to the definitions from van der Hofstad (2016), this reflects even a particular subset of smallworlds, namely ultra smallworlds.
Please note that the clustering coefficient takes values between 0 and 1.
In this paper, we have varied \(\sigma ^{neigh}\) and \(\phi \) between 4 and 6 to favor the emergence of local medium connected neighborhoods.
Please note that this procedure yields approximately the same number of connections in GRBM\(_\text {FixProb}\)/RBM\(_\text {FixProb}\) as in GXBM/XBM to ensure a fair comparison.
http://yann.lecun.com/exdb/mnist/, Last visit on October \(18\mathrm{th}\) 2015.
The average cluster coefficient was computed using the method proposed for bipartite graphs in Latapy et al. (2008).
Please note that a normal RGB image of 1000 by 1000 pixels has 3000000 dimensions.
References
Ackley, H., Hinton, E., & Sejnowski, J. (1985). A learning algorithm for Boltzmann machines. Cognitive Science, 9, 147–169.
Ammar, H. B., Mocanu, D. C., Taylor, M., Driessens, K., Tuyls, K., & Weiss, G. (2013). Automatically mapped transfer between reinforcement learning tasks via threeway restricted boltzmann machines. In H. Blockeel, K. Kersting, S. Nijssen, & F. Elezn (Eds.), Machine learning and knowledge discovery in databases (Vol. 8189, pp. 449–464). Lecture Notes in Computer Science Berlin: Springer. doi:10.1007/9783642409912_29.
Ba, J., & Caruana, R. (2014). Do deep nets really need to be deep? Advances in Neural Information Processing Systems, 27, 2654–2662.
Barabasi, A. L., & Albert, R. (1999). Emergence of scaling in random networks. Science, 286(5439), 509–512. doi:10.1126/science.286.5439.509.
Bengio, Y. (2009). Learning deep architectures for AI. Foundations and Trends in Machine Learning, 2(1), 1–127. doi:10.1561/2200000006.
Brgge, K., Fischer, A., & Igel, C. (2013). The flipthestate transition operator for restricted Boltzmann machines. Machine Learning, 93(1), 53–69. doi:10.1007/s1099401353903.
CarreiraPerpinan, M. A., & Hinton, G. E. (2005). On contrastive divergence learning. In 10th international workshop on artificial intelligence and statistics (AISTATS).
Clauset, A., Shalizi, C. R., & Newman, M. E. J. (2009). Powerlaw distributions in empirical data. SIAM Review, 51(4), 661–703. doi:10.1137/070710111.
Del Genio, C. I., Gross, T., & Bassler, K. E. (2011). All scalefree networks are sparse. Physical Review Letter, 107, 178701. doi:10.1103/PhysRevLett.107.178701.
Desjardins, G., Courville, A., Bengio, Y., Vincent, P., & Delalleau, O. (2010). Tempered Markov Chain Monte Carlo for training of restricted Boltzmann machines. In Y. W. Teh, & M. Titterington (Eds.), Proceedings of the thirteenth international conference on artificial intelligence and statistics (pp. 145–152), May 13–15, 2010. Sardinia: Chia Laguna Resort.
Dieleman, S., & Schrauwen, B. (2012). Accelerating sparse restricted Boltzmann machine training using nongaussianity measures. In Y. Bengiom, J. Bergstra, & Q. Le (Eds.) Proceedings of deep learning and unsupervised feature learning (p. 9).
Gehler, P. V., Holub, A. D., & Welling, M. (2006). The rate adapting poisson model for information retrieval and object recognition. In Proceedings of the 23rd international conference on machine learning, ACM, ICML ’06 (pp. 337–344). doi:10.1145/1143844.1143887.
Germain, M., Gregor, K., Murray, I., & Larochelle, H. (2015). MADE: Masked autoencoder for distribution estimation. In Proceedings of the 32nd international conference on machine learning, JMLR.org, JMLR Proceedings (Vol. 37, pp. 881–889).
Hagberg, A. A., Schult, D. A., & Swart, P. J. (2008). Exploring network structure, dynamics, and function using NetworkX. In Proceedings of the 7th python in science conference (SciPy2008) (pp. 11–15), Pasadena: CA USA.
Hakimi, S. L. (1962). On realizability of a set of integers as degrees of the vertices of a linear graph. I. Journal of Society for Industrial and Applied Mathematics, 10, 496–506.
Han, S., Pool, J., Tran, J., & Dally, W. (2015). Learning both weights and connections for efficient neural network. In C. Cortes, N. D. Lawrence, D. D. Lee, M. Sugiyama, & R. Garnett (Eds.), Advances in neural information processing systems (Vol. 28, pp. 1135–1143). Red Hook: Curran Associates Inc.
Hinton, G., (2012). A practical guide to training restricted Boltzmann machines. In Neural networks: Tricks of the trade (Vol. 7700, pp. 599–619). Lecture Notes in Computer Science. Berlin: Springer. doi:10.1007/9783642352898_32.
Hinton, G. E. (2002). Training products of experts by minimizing contrastive divergence. Neural Computation, 14(8), 1771–1800. doi:10.1162/089976602760128018.
Hinton, G. E., & Salakhutdinov, R. R. (2006). Reducing the dimensionality of data with neural networks. Science, 313(5786), 504–507. doi:10.1126/science.1127647.
Jones, N. (2014). Computer science: The learning machines. Nature, 505(7482), 146–148.
Krizhevsky, A., Sutskever, I., & Hinton, G. E. (2012). Imagenet classification with deep convolutional neural networks. Advances in Neural Information Processing Systems, 25, 1097–1105.
Larochelle, H., & Bengio, Y. (2008). Classification using discriminative restricted Boltzmann machines. In Proceedings of the 25th international conference on machine learning, ACM, ICML ’08 (pp. 536–543). doi:10.1145/1390156.1390224
Larochelle, H., & Murray, I. (2011). The neural autoregressive distribution estimator. In AISTATS, JMLR.org, JMLR Proceedings (Vol. 15, pp 29–37).
Latapy, M., Magnien, C., & Vecchio, N. D. (2008). Basic notions for the analysis of large twomode networks. Social Networks, 30(1), 31–48. doi:10.1016/j.socnet.2007.04.006.
Lee, H., Ekanadham, C., & Ng, A. Y. (2008). Sparse deep belief net model for visual area v2. In J. Platt, D. Koller, Y. Singer, & S. Roweis (Eds.), Advances in Neural Information Processing System (Vol. 20, pp. 873–880). Red Hook: Curran Associates Inc.
Lee, H., Pham, P., Largman, Y., & Ng, A. Y. (2009). Unsupervised feature learning for audio classification using convolutional deep belief networks. Advances in Neural Information Processing Systems, 22, 1096–1104.
Luo, H., Shen, R., Niu, C., & Ullrich, C. (2011). Sparse group restricted boltzmann machines. In W. Burgard & D. Roth (Eds.), AAAI. Menlo Park: AAAI Press.
Marlin, B. M., Swersky, K., Chen, B., & de Freitas, N. (2010). Inductive principles for restricted boltzmann machine learning. In AISTATS, JMLR.org, JMLR Proceedings (Vol. 9, pp. 509–516)
Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A. A., Veness, J., Bellemare, M. G., et al. (2015). Humanlevel control through deep reinforcement learning. Nature, 518(7540), 529–533. doi:10.1038/nature14236.
Mocanu, D. C., Ammar, H. B., Lowet, D., Driessens, K., Liotta, A., Weiss, G., et al. (2015). Factored four way conditional restricted boltzmann machines for activity recognition. Pattern Recognition Letters, 66, 100–108. doi:10.1016/j.patrec.2015.01.013. Pattern Recognition in Human Computer Interaction.
Osogami, T., & Otsuka, M. (2014). Restricted Boltzmann machines modeling human choice. Advances in Neural Information Processing Systems, 27, 73–81.
Pessoa, L. (2014). Understanding brain networks and brain organization. Physics of Life Reviews, 11(3), 400–435. doi:10.1016/j.plrev.2014.03.005.
Ranzato, M. A., lan Boureau, Y., & Cun, Y. L. (2008). Sparse feature learning for deep belief networks. In J. Platt, D. Koller, Y. Singer, & S. Roweis (Eds.), Advances in neural information processing systems (Vol. 20, pp. 1185–1192). Red Hook: Curran Associates Inc.
Salakhutdinov, R., & Murray, I. (2008). On the quantitative analysis of deep belief networks. In Proceedings of the international conference on machine learning (pp. 872–879).
Salakhutdinov, R., Mnih, A., & Hinton, G. (2007). Restricted boltzmann machines for collaborative filtering. In Proceedings of the 24th international conference on machine learning, ACM, ICML ’07 (pp 791–798). doi:10.1145/1273496.1273596
Smolensky, P. (1987). Information processing in dynamical systems: Foundations of harmony theory. In D. E. Rumelhart, J. L. McClelland, et al. (Eds.), Parallel distributed processing (Vol. 1, pp. 194–281). Cambridge: MIT Press.
Srivastava, N., & Salakhutdinov, R. R. (2012). Multimodal learning with deep Boltzmann machines. In F. Pereira, C. Burges, L. Bottou, & K. Weinberger (Eds.), Advances in neural information processing systems (Vol. 25, pp. 2222–2230). Red Hook: Curran Associates Inc.
Strogatz, S. H. (2001). Exploring complex networks. Nature, 410(6825), 268–276.
Swersky, K., Tarlow, D., Sutskever, I., Salakhutdinov, R., Zemel, R. S., & Adams, R. P. (2012). Cardinality restricted boltzmann machines. In NIPS (pp. 3302–3310).
Tieleman, T. (2008). Training restricted boltzmann machines using approximations to the likelihood gradient. In Proceedings of the 25th international conference on machine learning, ICML ’08 (pp. 1064–1071), ACM, New York, NY, USA. doi:10.1145/1390156.1390290
Tieleman, T., & Hinton, G. (2009). Using fast weights to improve persistent contrastive divergence. In Proceedings of the 26th annual international conference on machine learning, ICML ’09 (pp. 1033–1040), ACM, New York, NY, USA. doi:10.1145/1553374.1553506
van der Hofstad, R. (2016). Random graphs and complex networks (Vol. i). http://www.win.tue.nl/~rhofstad/NotesRGCN
Wan, C., Jin, X., Ding, G., & Shen, D. (2015) Gaussian cardinality restricted boltzmann machines. In Twentyninth AAAI conference on artificial intelligence.
Watts, D. J., & Strogatz, S. H. (1998). Collective dynamics of ’smallworld’ networks. Nature, 393, 440–442.
Welling, M., Rosenzvi, M., & Hinton, G. E. (2005). Exponential family harmoniums with an application to information retrieval. In L. Saul, Y. Weiss, & L. Bottou (Eds.), Advances in neural information processing systems (Vol. 17, pp. 1481–1488). Cambridge: MIT Press.
Yosinski, J., & Yosinski and Lipson, H. (2012). Visually debugging restricted Boltzmann machine training with a 3d example. In Representation learning workshop, 29th international conference on machine learning.
Zhou, F., Claire, Q., & King, R. (2014). Predicting the geographical origin of music. In 2014 IEEE international conference on data mining (ICDM) (pp. 1115–1120). doi:10.1109/ICDM.2014.73
Acknowledgments
This research has been partly funded by the European Union’s Horizon 2020 project INTERIoT (Grant Number 687283), and by the NL Enterprise Agency under the TKI SGBEMS project of Dutch Top Sector.
Author information
Authors and Affiliations
Corresponding author
Additional information
Editors: Thomas Gaertner, Mirco Nanni, Andrea Passerini, and Celine Robardet.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Cite this article
Mocanu, D.C., Mocanu, E., Nguyen, P.H. et al. A topological insight into restricted Boltzmann machines. Mach Learn 104, 243–270 (2016). https://doi.org/10.1007/s109940165570z
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s109940165570z
Keywords
 Deep learning
 Sparse restricted Boltzmann machines
 Complex networks
 Scalefree networks
 Smallworld networks