Research Article | Open Access

# Topologically Ordered Feature Extraction Based on Sparse Group Restricted Boltzmann Machines

**Academic Editor:**Panos Liatsis

#### Abstract

How to extract topologically ordered features efficiently from high-dimensional data is an important problem of unsupervised feature learning domains for deep learning. To address this problem, we propose a new type of regularization for Restricted Boltzmann Machines (RBMs). Adding two extra terms in the log-likelihood function to penalize the group weights and topologically ordered factors, this type of regularization extracts topologically ordered features based on sparse group Restricted Boltzmann Machines (SGRBMs). Therefore, it encourages an RBM to learn a much smoother probability distribution because its formulations turn out to be a combination of the group weight-decay and topologically ordered factor regularizations. We apply this proposed regularization scheme to image datasets of natural images and Flying Apsara images in the Dunhuang Grotto Murals at four different historical periods. The experimental results demonstrate that the combination of these two extra terms in the log-likelihood function helps to extract more discriminative features with much sparser and more aggregative hidden activation probabilities.

#### 1. Introduction

Restricted Boltzmann Machines (RBMs) [1] are a type of product of experts model [2] based on Boltzmann Machines [3] but with a complete bipartite interaction graph. In general, RBMs, which are used as generative models to simulate input distributions of binary data [4], are viewed as an effective feature-representation approach for extracting structured information from input data. They have received much attention recently and have been successfully applied in various application domains, such as dimensionality reduction [5], object recognition [6], topic modeling [7], and feature learning [8]. In addition, RBMs have attracted much attention as building blocks for the multilayer learning systems (e.g., Deep Belief Networks (DBNs), Deep Boltzmann Machines (DBMs)), and variants and extensions of RBMs have a great many applications in a wide range of feature learning and pattern recognition tasks.

Due to the arbitrary connectivity of Boltzmann machines, they are too slow to be practical, and in order to obtain efficient and exact results, RBMs have the restrictions that there are no visible-visible or hidden-hidden connections, which leads to the obvious advantage that inferences in the RBMs are much easier than in Boltzmann Machines [9]. Therefore, the hidden units are conditionally independent and we may generate a more powerful learning model [10]. Lee et al. [11] proposed sparse RBMs (SRBMs) by pointing out that RBMs tend to learn distributed and nonsparse representations as the number of hidden units is increased; accordingly, they added a regularization term that penalized a deviation of the expected activation with a low level to ensure that the hidden units would be sparsely activated. Moreover, in order to group similar activations of the hidden units and capture their local dependencies, Luo et al. [12] proposed sparse group RBMs (SGRBMs) using a novel regularization of the activation probabilities of the hidden units in RBMs. What SRBMs and SGRBMs have in common is that they have adopted sparsity to promote regularization, making them powerful enough to represent complicated distributions.

By introducing the regularizer into the activation probabilities of the hidden units, the SGRBMs have the following two properties: first, this model encourages few groups to be active when given observed data (this property yields sparsity at group level), and second, it results in only a few hidden units being active in a group (this property yields sparsity within the group). However, they did not consider overfitting problems, which lack corresponding strategies for controlling the reconstruction complexity of the weight matrix. In addition, they did not take into account the fact that all the extracted features in the hidden units are not topologically ordered (i.e., similar features are grouped together while they do not simultaneously discard group sparsity), and it is essential for a learning machine to obtain structured information from the input data. In 2002, Welling et al. [13] proposed a novel learning sparse topographic representation with products of Student -distributions and found that if the Student -distribution is used to model the combined outputs of sets of neutrally adjacent filters, then the orientation, spatial frequency, and location of the filters change smoothly across the topographic map. Later, Goh et al. [14] proposed a method for regularizing RBMs during training to obtain features that are sparse and topographically organized. The features learned are then Gabor-like and demonstrate a coding for orientation, spatial position, frequency, and color that vary smoothly with the topography of the feature map. For the purpose of efficiently extracting invariant features with group sparsity from high-dimensional data, in this paper, firstly we adopted a weight-decay strategy [15, 16] at group level based on SGRBMs, and secondly, by adding an extra term to penalize the topologically ordered factors in the log-likelihood function, the topologically ordered features at group level can be obtained.

The remaining sections of this paper are organized as follows. In Section 2, RBMs and Contrastive Divergence algorithms for RBM training are described in brief. In Section 3, a nontopologically ordered feature extraction approach is proposed to obtain sparse but not topologically ordered features between groups from the input data. In Section 4, a topologically ordered feature extraction approach is proposed to obtain structured information (i.e., sparse and topologically ordered features between the overlapping groups) from the input data. In Section 5, experimental results with two different datasets (namely, natural images and Flying Apsara images in the Dunhuang Grotto Murals) are shown to validate the proposed approach. Finally, the conclusions are in Section 6.

#### 2. Restricted Boltzmann Machines and Contrastive Divergence

RBMs are a particular form of the Markov Random Field (MRF) model and are regarded as an undirected generative model which uses a layer of binary hidden units to model a distribution over binary visible units [17]. Suppose an RBM consists of visible units representing the input data and hidden units to capture the features of the input data. The joint probability distribution is given by the Gibbs distribution with the energy function [17, 18]:where is the matrix of weights and and are vectors which represent the visible and hidden biases, respectively. All these are referred to as the RBM parameters , and is the energy function and is a corresponding normalized constant. Therefore, the marginal distribution of visible variables becomes

As the hidden units are independent given the states of the visible units and vice versa, when given the observed data, the conditional probabilities and conditional distributions of the hidden units arewhere is the th column of , which is a vector that represents the connection weights between the th hidden unit and all visible units, and is the sigmoid activation function. Thus, the marginal distribution of the visible variables actually is a model of product of experts [2, 12]:Equation (4) deduces that all these hidden units for the individual components of the given data vector are combined multiplicatively and will contribute probabilities according to the activation probabilities. If given a data sample, one specific hidden unit will be activated with a high probability, and the hidden unit is responsible for representing the data sample. If more data in the training data set activates a hidden unit with a higher probability, the hidden unit’s feature will be less discriminative. Thus it is sometimes necessary to introduce sparsity at the hidden layer of an RBM [11, 19, 20].

For a training example , training an RBM is the same as modeling the marginal distribution of the visible units. A common practice is to adopt the log-likelihood gradient approach [16, 18] to maximizing the marginal distribution , which aims to generate with the largest probability. Using gradient descent approach can solve this problem:The second term of (5) is intractable because we cannot obtain any information about the marginal distribution . In order to solve this problem, Hinton et al. [21] proposed Contrastive Divergence (CD) learning, which has become a standard way to train RBMs. The -step contrastive divergence learning () is in two steps: first, the Gibbs chain is initialized with the training example of the training set. Second, the sample is yielded after steps, and each step consists of sampling from and subsequently sampling from . According to the general Markov Chain Monte Carlo (MCMC) theory, we know that when , the -step contrastive divergence learning algorithm converges to the second term of (5) and becomes only visible in the proof of Bengio and Delalleau [22]. However, Hinton [16] pointed out that when initializing with the training example of the training set, running one-step () Gibbs sampling approximates this term in the log-likelihood gradient relatively well. Therefore, (5) can be approximated asThus, the iterative update process of the th column from the weight matrix for the training example is represented bywhere is the learning rate. The first term of (7) decreases the energy of [23]; at the same time, this term also guarantees that unit is more likely to be activated when the hidden unit observes again; this means that the hidden units are learning to represent [12]. In the next section, we use a weight-decay strategy at group level for SGRBMs to capture features with group sparsity from the input data.

#### 3. Nontopologically Ordered Feature Extraction Based on Sparse Group RBMs

In the unsupervised learning process, some of the hidden units may extract similar features if there is little difference between their corresponding weight vectors . This homogenization problem can be obvious and serious if the number of the hidden units is increased. To alleviate this problem, Lee et al. [11] introduced SRBMs and Luo et al. [12] introduced SGRBMs to remit statistical dependencies between all of the hidden units when adding a penalty term. SRBMs have been popular due to the fact that an RBM with a low average hidden activation probability is better at extracting discriminative features than nonregularized RBMs [8, 24]. This is especially the case in Luo et al. [12], who divided the hidden units equally into nonoverlapping groups to restrain the dependencies within these groups and penalized the overall activation level of a group. More discriminative features are learned when SGRBMs are applied to deep learning systems for classification tasks. However, Luo et al. [12] did not consider overfitting problems and did not propose any strategies for controlling the reconstruction complexity of the weight matrix. Thus, to equilibrate the reconstruction error (i.e., the learning accuracy of specific training samples) and reconstruction complexity (generalization ability), we have used a weight-decay strategy at group level based on SGRBMs to capture features with sparse grouping of the input data.

For an RBM with hidden units, let denote the set of all indices of the hidden units. The th group is denoted by , where , . Suppose all groups are nonoverlapping and of equal size [12] (see Figure 1). Given a grouping and a sample , the th group norm is given bywhere is the (Euclidean) norm of the vector comprising the activation probabilities , which are considered as the overall activation level of the th group. Given all the group norms , the mixed norm is

The mixed norm of the grouping weight vectors is shown byIn fact, the mixed norm is considered as the overall weight strength level of the grouping.

We add these two regularizers ((9) and (11)) to the log-likelihood of the training examples. Thus, given the training set comprising examples, we need to solve the following optimization problem:where and are two regularization constants. The second term increases the sparsity of the hidden units at group level, and the third term decreases the reconstruction complexity of this model. We apply the contrastive divergence update rule (see (7)) to solve (12), and it is followed by one step of gradient ascent by using the gradient of the regularization terms.

By introducing these two regularizers, the iterative process to solve the optimal parameters (7) is updated as follows:where is the - sampling from , and we assume the th hidden unit belongs to the th group . The last step in (13) is derived from . Since the second and third terms in (12) control the sparsity of activation probabilities of the hidden units, then using (13), we can regard the hidden layer of RBMs as the nontopological feature extractor to capture features with group sparsity from the training data, abbreviated to NTOSGRBMs. In the next section, we obtain topologically ordered features from the input data at group level by adding an extra term to the mixed norm of the group weights based on SGRBMs and by adding an extra term to penalize the topologically ordered factors in the log-likelihood function.

#### 4. Topologically Ordered Feature Extraction Based on Sparse Group RBMs

The approach of the sparsity-based feature extraction approaches is to employ regularizers to induce sparsity during discriminative feature representation [25, 26]. According to Luo et al. [12], the mixed norm encourages sparsity at group level; however, it does not contain any prior information about possible groups of covariates that we may want to select jointly [27, 28]. From the SGRBMs, we can learn a set of features with group sparsity that is useful for representing the input data; however, drawing inspiration from the human brain, we would like to learn a set of features that have both with group sparsity and topologically ordered in some manner. Here,* topologically ordered method* means that similar features are grouped together while retaining group sparsity. The aim of this constraint for hidden units is to group adjacent features together in the smoothed penalty. Instead of keeping all groups being nonoverlapping, therefore, we retain all of the overlapping groups. Then, suppose the th group is denoted by , where , , and . Also, suppose that all of these overlapping groups are of the same size; then all overlapping parts are of the same size, and each hidden unit belongs to two neighboring groups (see Figure 2).

Since represents the overall activation level of hidden units for all groups, then (12) becomes

In fact, since we actually minimize the overall weight strength level of the grouping, the third term in (14) ensures that only a few hidden units can be activated in a group. From the perspective of information theory, the entropy of the distribution of the conditional probabilities for all hidden units in a group is relatively low. For the th group in the hidden layer, the entropy of the conditional probabilities’ distribution is defined as

For the neighboring two groups, the th group and th group at the hidden layer, we can define the topologically ordered factor (TOF) between these neighboring two groups aswhere is the count function to obtain the number of elements in a set. The structured features in the hidden units are topologically well-ordered if is close to zero. Since one important research direction in using the stochastic method (i.e., contrastive divergence [2] and approximate maximum-likelihood [29]) for RBMs is to design a regularization term [30], it is common to use weight-decay that regularizes the growth of parameters to avoid overfitting and stabilize learning. Therefore, in this paper, we have added an extra term to penalize the topologically ordered factors, so that the topologically ordered feature extraction based on SGRBMs (TOSGRBMs) can be extended by these two extra regularizers. Thus, given training data , we need to solve the following optimization problem:This type of regularization can be seen as the combination of the group weight-decay and topologically ordered factors regularization. To address this problem, it is possible to use the alternating direction method of gradient ascent. The partial derivative of the first and third terms is shown in (13), and this problem may be turned into solving the partial derivative term . Suppose ; thus and ; then

The objective function in (17) is optimized when using the iterated method described. In the above discussion, the form of (13) is not changed, but one significant difference is that we assume the th hidden unit belongs to the th group and its neighboring group: the th group . Thus, the iterative formula for the optimal solution of parameters of RBM iswhere . The first term of (17) represents the true distribution of the input data, and maximizing it in fact minimizes the reconstruction error of RBMs. The third term of (17) represents the group sparsity of the hidden layer, and minimizing this penalty term is equivalent to minimizing the reconstruction complexity of RBMs. In addition, in order to obtain topologically ordered features from these features with group sparsity, we add the second term of (17) to penalize the topologically ordered factors.

The objective function in (17) is actually an optimization problem; however, it is not convex. In principle, we can apply gradient descent to solve this problem; however, computing the gradient of the log-likelihood term is expensive. Fortunately, the contrastive divergence learning algorithm gives an efficient approximation to the gradient of the log-likelihood [2]. Building upon this, in each iteration we can apply the contrastive divergence update rule, followed by one step of gradient descent using the gradient of the regularization terms [11].

#### 5. Experimental Results

In this section, we compared the results of the proposed nontopologically ordered and topologically ordered feature extraction approaches based on SGRBMs. First, we applied the two approaches to model patches of natural images. Then, we applied them to analyze the structured features of Flying Apsara images from the Dunhuang Grotto Murals at four different historical periods.

##### 5.1. Modeling Patches of Natural Images

The training data consists of 20,000 patches , randomly extracted from a standard set of ten whitened images. All the patches were divided into minibatches, each containing 256 patches, and updated the weights of each minibatch (total batches = 2,000).

We trained a nontopologically ordered SGRBM with 20,000 real-valued visible units and 256 hidden units which were divided into 64 uniform nonoverlapping groups containing four hidden units each. The learning rate was set to 0.1 [16], and the regularization constants and were empirically set to 0.1 and 0.5, respectively. The learned features are shown in Figure 3(a). For comparison, we also trained a topologically ordered SGRBM with 256 hidden units. The learned features are shown in Figure 3(b). Some features are extracted and localized Gabor-like edge detectors in different positions and orientations; these results are like those in [13, 14]. Since the hidden units within a group compete with each other in the modeling patches, each hidden unit in the nontopologically ordered SGRBMs focused on modeling more subtle patterns in the training data. As a result, the features learned with the topologically ordered SGRBMs are more aggregative at group level than those learned with the nontopologically ordered SGRBMs. Moreover, the learned features by TOSGRBMs shown in Figure 3 have an enforced a topological order, where the location, orientation, and frequency of the Gabor-like filters all change smoothly. In conclusion, from the perspective of the invariant feature learning, the topologically ordered feature extraction approach facilitate the training of the whole network to extract more discriminative features at the hidden layer.

**(a) Nontopologically ordered features**

**(b) Topologically ordered features**

We also compared our results with the standard SRBMs and SGRBMs. With the same parameter settings and the same number of iterations (=10^{4}), Figure 4 shows that the SRBMs, SGRBMs, NTOSGRBMs, and TOSGRBMs extracted Gabor-like filter features; however, SRBMs and SGRBMs have many more redundant feature patches than NTOSGRBMs and TOSGRBMs. Moreover, since the grouped features are significant in SGRBMs, NTOSGRBMs, and TOSGRBMs, the features learned by the SGRBMs, NTOSGRBMs, and TOSGRBMs are more localized than those learned by the SRBMs.

**(a) Minibatches**

**(b) SRBMs**

**(c) SGRBMs**

**(d) NTOSGRBMs**

**(e) TOSGRBMs**

In addition, we use Hoyer’s sparseness measure [31] to determine the sparse representations learned by the SRBMs, SGRBMs, NTOSGRBMs, and TOSGRBMs. This measure is in the interval and on a normalized scale. Figure 5 shows the activation probabilities of the hidden units that were computed using the regular SRBMs, SGRBMs, NTOSGRBMs, and TOSGRBMs. It can be seen that the representations learned by the SGRBMs were much more sparse than for the other three models, although NTOSGRBMs and TOSGRBMs, with similar activation probabilities of hidden units, learned similar representations (close sparseness values) and learned much more sparse representations than the SRBMs, but much less sparse representations than the SGRBMs.

**(a) SRBMs**

**(b) SGRBMs**

**(c) NTOSGRBMs**

**(d) TOSGRBMs**

##### 5.2. Modeling Patches of Flying Apsaras Images in the Dunhuang Grotto Murals

The image dataset of the Flying Apsaras in the Dunhuang Grotto Murals published in Zheng and Tai’s book [32] contains 300 images. These images cover four historical periods: the early stage, the developing stage, the flourishing stage, and the terminal stage [33]. In the present study, as an example, the training data consisted of 20,000 randomly selected image patches of the Flying Apsara images. These patches were randomly extracted from a standard set of ten fine-art paintings of the Flying Apsaras (Figure 6). Features from these images covering the four historical periods were exacted using both nontopologically ordered and topologically ordered SGRBMs. In addition, the parameters settings were the same as in the previous subsection.

**(a)**

**(b)**

In Figure 7, we see that both approaches show pronounced advantages in extracting discriminative features at group level, although the features learned by the topologically ordered SGRBMs were more aggregative than the nontopologically ordered SGRBMs. Moreover, since there is structural similarity between the features of the Flying Apsara images, their representations varied smoothly compared to the transformations and invariance achieved using TOSGRBMs. The features learned by TOSGRBMs (Figure 7) had enforced topological order, where the location, orientation, and frequency of the Gabor-like filters all change smoothly. It is concluded that, when using topologically ordered feature extraction based on SGRBMs, feature selection performed well because of the sparse and aggregative features at group level.

**(a) Nontopologically ordered features**

**(b) Topologically ordered features**

Taking the early stage of the Flying Apsara images as an example, we also compared our learned features with those of the standard SRBMs and SGRBMs. With the same parameter settings and the same number of iterations (=10^{4}), Figure 8 shows that the grouped features are significant in SGRBMs, NTOSGRBMs, and TOSGRBMs. The features learned using the SGRBMs, NTOSGRBMs, and TOSGRBMs are more localized than those for the SRBMs. Moreover, the sparse features learned with SRBMs and NTOSGRBMs did not group similar feature patches as an aggregative representation. Both SGRBMs and TOSGRBMs learned to group similar features without discarding group sparsity; however, the sparse features learned by TOSGRBMs not only grouped similar feature patches as an aggregative representation but also retained similar feature patches in topological order at group level.

**(a) Minibatches**

**(b) SRBMs**

**(c) SGRBMs**

**(d) NTOSGRBMs**

**(e) TOSGRBMs**

To measure the sparsity of the weight matrix quantitatively, we used a sparsity indicator based on the relationship between the and norms. In accordance with Cauchy Inequality, . The indicator takes a value of 1 if and only if contains only one single nonzero component and 0 if and only if all elements are equal (but not equal to zero) and interpolates smoothly between the two extremes [31]. The Hoyer measure is a normalized version of the measure. It approximates the measure but, because it is not as flat as the measure, it has a gradient that can be used in optimization problems [34].

Figures 9–11 list the different performances when applied to the Flying Apsara images for the four historical periods. Those show the average reconstruction error, the average topologically ordered factor, and the average sparsity of the weight matrix from five individual experiments with different random initializations for both nontopologically ordered and topologically ordered SGRBMs and different numbers of iterations (, , , , , , , and ). These results show that TOSGRBMs not only generate the topologically ordered features well (see Figures 7, 8, and 10) but also generate representative features well. Since the average sparsity of weight matrix is large (see Figure 11), TOSGRBMs give the hidden units low correlation when representing structured features of the input data, and according to the tendency of the average reconstruction errors (Figure 9), it can avoid over-fitting learning simultaneously. The topologically ordered SGRBMs approach achieved the best discriminative feature extraction and produced the best trade-off between reconstruction error and complexity. The topologically ordered SGRBMs have lower average topologically ordered factors, which indicate that the proposed topologically ordered SGRBMs decrease the similarities between extracted features and order them well topologically, because of the penalty on the topologically ordered factors of all groups.

#### 6. Conclusions

For the purpose of extracting topologically ordered features efficiently from high-dimensional data, firstly we used a weight-decay strategy at group level based on SGRBMs to capture features with group sparsity from the input data. Secondly, by adding an extra term to penalize the topologically ordered factors in the log-likelihood function, we obtain topologically ordered features at group level. Experimental results on the image datasets of both natural images and the Flying Apsara images from the Dunhuang Grotto Murals at four different historical periods demonstrate that the combination of these two extra terms in the log-likelihood function helps to extract better discriminative features with much sparser and more aggregative hidden activation probabilities. In conclusion, in our experiments the topologically ordered SGRBMs showed markedly prominent sparsity of weight matrix, discriminative features at the hidden layer, and sparse feature representations. Topologically ordered SGRBMs were therefore found to be superior to nontopologically ordered SGRBMs for those reasons.

#### Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

#### Acknowledgments

This work was supported in part by the National Key Basic Research Program of China (no. 2012CB725303), the National Science Foundation of China (no. 61170202, no. 41371420, and no. 61203154), and the Fundamental Research Funds for the Central Universities (no. 2013-YB-003). The authors would like to thank the editor and the anonymous reviewers for their valuable comments.

#### References

- Y. Freund and D. Haussler, “Unsupervised learning of distributions on binary vectors using two layer networks,” Tech. Rep., Computer Research Laboratory, University of California, Santa Cruz, Calif, USA, 1994. View at: Google Scholar
- G. E. Hinton, “Training products of experts by minimizing contrastive divergence,”
*Neural Computation*, vol. 14, no. 8, pp. 1771–1800, 2002. View at: Publisher Site | Google Scholar - D. H. Ackley, G. E. Hinton, and T. J. Sejnowski, “A learning algorithm for boltzmann machines,”
*Cognitive Science*, vol. 9, no. 1, pp. 147–169, 1985. View at: Publisher Site | Google Scholar - G. F. Montúfar, J. Rauh, and N. Ay, “Expressive power and approximation errors of restricted Boltzmann machines,” in
*Proceedings of the Advances in Neural Information Processing Systems (NIPS '11)*, pp. 415–423, Granada, Spain, December 2011. View at: Google Scholar - G. E. Hinton and R. R. Salakhutdinov, “Reducing the dimensionality of data with neural networks,”
*Science*, vol. 313, no. 5786, pp. 504–507, 2006. View at: Publisher Site | Google Scholar | MathSciNet - H. Lee, R. Grosse, R. Ranganath, and A. Y. Ng, “Convolutional deep belief networks for scalable unsupervised learning of hierarchical representations,” in
*Proceedings of the 26th International Conference On Machine Learning (ICML '09)*, pp. 609–616, June 2009. View at: Google Scholar - G. E. Hinton and R. Salakhutdinov, “Replicated softmax: an undirected topic model,” in
*Proceedings of the Advances in Neural Information Processing Systems (NIPS '09)*, pp. 1607–1614, 2009. View at: Google Scholar - A. Coates, A. Y. Ng, and H. Lee, “An analysis of single-layer networks in unsupervised feature learning,” in
*Proceedings of the 14th International Conference on Artificial Intelligence and Statistics (AISTATS '11)*, pp. 215–223, Fort Lauderdale, Fla, USA, April 2011. View at: Google Scholar - M. Yasuda and K. Tanaka, “Approximate learning algorithm for restricted Boltzmann machines,” in
*Proceedings of the International Conference on Computational Intelligence for Modelling Control & Automation*, pp. 692–697, IEEE, Vienna, Austria, December 2008. View at: Publisher Site | Google Scholar - P. Garrigues and B. Olshausen, “Learning horizontal connections in a sparse coding model of natural images,” in
*Proceedings of the Advances in Neural Information Processing Systems (NIPS '08)*, vol. 20, MIT Press, Cambridge, Mass, USA, 2008. View at: Google Scholar - H. Lee, C. Ekanadham, and A. Y. Ng, “Sparse deep belief net model for visual area v2,” in
*Proceedings of the 21st Annual Conference on Neural Information Processing Systems (NIPS '07)*, pp. 873–880, December 2007. View at: Google Scholar - H. Luo, R. Shen, C. Niu, and C. Ullrich, “Sparse group restricted Boltzmann machines,” in
*Proceedings of the 25th AAAI Conference on Artificial Intelligence*, pp. 429–434, San Francisco, Calif, USA, August 2011. View at: Google Scholar - M. Welling, G. Hinton, and S. Osindero, “Learning sparse topographic representations with products of student-t distributions,” in
*Advances in Neural Information Processing Systems (NIPS)*, pp. 1359–1366, MIT Press, 2002. View at: Google Scholar - H. Goh, L. Kusmierz, J.-H. Lim, N. Thome, and M. Cord, “Learning invariant color features with sparse topographic restricted Boltzmann machines,” in
*Proceedings of the 18th IEEE International Conference on Image Processing (ICIP '11)*, pp. 1241–1244, IEEE, Brussels, Belgium, September 2011. View at: Publisher Site | Google Scholar - S. Bengio, F. Pereira, Y. Singer, and D. Strelow, “Group sparse coding,” in
*Proceedings of the Advances in Neural Information Processing Systems (NIPS '09)*, pp. 82–89, 2009. View at: Google Scholar - G. E. Hinton, “A practical guide to training restricted Boltzmann machines,”
*Momentum*, vol. 9, no. 1, p. 926, 2010. View at: Google Scholar - J. Martens, A. Chattopadhya, T. Pitassi, and R. Zemel, “On the representational efficiency of restricted Boltzmann machines,” in
*Proceedings of the Advances in Neural Information Processing Systems (NIPS '13)*, pp. 2877–2885, December 2013. View at: Google Scholar - A. Fischer and C. Igel, “Training restricted Boltzmann machines: an introduction,”
*Pattern Recognition*, vol. 47, no. 1, pp. 25–39, 2014. View at: Publisher Site | Google Scholar - H. Larochelle and Y. Bengio, “Classification using discriminative restricted Boltzmann machines,” in
*Proceedings of the 25th International Conference on Machine Learning (ICML '08)*, pp. 536–543, ACM, Helsinki, Finland, July 2008. View at: Publisher Site | Google Scholar - R. Salakhutdinov and H. Larochelle, “Efficient learning of deep Boltzmann machines,” in
*Proceedings of the 13th International Conference on Artificial Intelligence and Statistics*, vol. 6, Sardinia, Italy, May 2010. View at: Google Scholar - G. E. Hinton, S. Osindero, and Y.-W. Teh, “A fast learning algorithm for deep belief nets,”
*Neural Computation*, vol. 18, no. 7, pp. 1527–1554, 2006. View at: Publisher Site | Google Scholar | Zentralblatt MATH - Y. Bengio and O. Delalleau, “Justifying and generalizing contrastive divergence,”
*Neural Computation*, vol. 21, no. 6, pp. 1601–1621, 2009. View at: Publisher Site | Google Scholar | MathSciNet - Y. Bengio, “Learning deep architectures for AI,”
*Foundations and Trends in Machine Learning*, vol. 2, no. 1, pp. 1–27, 2009. View at: Publisher Site | Google Scholar | Zentralblatt MATH - J. Ngiam, A. Khosla, M. Kim, J. Nam, H. Lee, and A. Y. Ng, “Multimodal deep learning,” in
*Proceedings of the 28th International Conference on Machine Learning (ICML '11)*, pp. 689–696, Bellevue, Wash, USA, June-July 2011. View at: Google Scholar - Q. Zhou, L. Zhang, and L. Ma, “Learning topographic sparse coding through similarity function,” in
*Proceedings of the 4th International Conference on Natural Computation (ICNC '08)*, pp. 241–245, IEEE, Jinan, China, October 2008. View at: Publisher Site | Google Scholar - F. Wu, Y. Yuan, Y. Rui, S. Yan, and Y. Zhuang, “Annotating web images using NOVA: non-convex group spArsity,” in
*Proceedings of the 20th ACM International Conference on Multimedia (MM '12)*, pp. 509–518, ACM, Nara, Japan, November 2012. View at: Publisher Site | Google Scholar - L. Jacob, G. Obozinski, and J.-P. Vert, “Group lasso with overlap and graph lasso,” in
*Proceedings of the 26th International Conference On Machine Learning (ICML '09)*, pp. 433–440, June 2009. View at: Google Scholar - J. Mairal, R. Jenatton, G. Obozinski, and F. Bach, “Learning hierarchical and topographic dictionaries with structured sparsity,” in
*Wavelets and Sparsity XIV*, Proceedings of SPIE, 81381P, p. 13, International Society for Optics and Photonics, 2011. View at: Google Scholar - T. Tieleman, “Training restricted boltzmann machines using approximations to the likelihood gradient,” in
*Proceedings of the 25th International Conference on Machine Learning*, pp. 1064–1071, July 2008. View at: Google Scholar - K. Cho, A. Ilin, and T. Raiko, “Tikhonov-type regularization for restricted Boltzmann machines,” in
*Artificial Neural Networks and Machine Learning—ICANN 2012*, vol. 7552 of*Lecture Notes in Computer Science*, pp. 81–88, Springer, Berlin, Germany, 2012. View at: Publisher Site | Google Scholar - P. O. Hoyer, “Non-negative matrix factorization with sparsity constraints,”
*Journal of Machine Learning Research*, vol. 5, pp. 1457–1469, 2004. View at: Google Scholar - R. Zheng and J. Tai,
*Collected Edition of Flying Apsaras in the Dunhuang Grotto Murals, China*, Commercial Press, Hong Kong, 2002 (Chinese). - Z. Chen, S. Xiong, Z. Fang, Q. Li, B. Wang, and Q. Zou, “A kernel support vector machine-based feature selection approach for recognizing Flying Apsaras' streamers in the Dunhuang Grotto Murals, China,”
*Pattern Recognition Letters*, vol. 49, pp. 107–113, 2014. View at: Publisher Site | Google Scholar - N. Hurley and S. Rickard, “Comparing measures of sparsity,”
*IEEE Transactions on Information Theory*, vol. 55, no. 10, pp. 4723–4741, 2009. View at: Publisher Site | Google Scholar | MathSciNet

#### Copyright

Copyright © 2015 Zhong Chen et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.