Computational Intelligence and Neuroscience

Volume 2008 (2008), Article ID 361705, 10 pages

http://dx.doi.org/10.1155/2008/361705

## Nonnegative Matrix Factorization with Gaussian Process Priors

^{1}Department of Informatics and Mathematical Modelling, Technical University of Denmark, Richard Petersens Plads, DTU-Building 321, 2800 Lyngby, Denmark^{2}Department of Electronic Systems, Aalborg University, Niels Jernes Vej 12, 9220 Aalborg Ø., Denmark

Received 31 October 2007; Revised 16 January 2008; Accepted 10 February 2008

Academic Editor: Wenwu Wang

Copyright © 2008 Mikkel N. Schmidt and Hans Laurberg. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

We present a general method for including prior knowledge in a nonnegative matrix factorization (NMF), based on Gaussian process priors. We assume that the nonnegative factors in the NMF are linked by a strictly increasing function to an underlying Gaussian process specified by its covariance function. This allows us to find NMF decompositions that agree with our prior knowledge of the distribution of the factors, such as sparseness, smoothness, and symmetries. The method is demonstrated with an example from chemical shift brain imaging.

#### 1. Introduction

Nonnegative matrix factorization (NMF) [1, 2] is a recent method for factorizing a matrix as the product of two matrices, in which all elements are nonnegative. NMF has found widespread application in many different areas including pattern recognition [3], clustering [4], dimensionality reduction [5], and spectral analysis [6, 7]. Many physical signals, such as pixel intensities, amplitude spectra, and occurrence counts, are naturally represented by nonnegative numbers. In the analysis of mixtures of such data, nonnegativity of the individual components is a reasonable constraint. Recently, a very simple algorithm [8] for computing the NMF was introduced. This has initiated much research aimed at developing more robust and efficient algorithms.

Efforts have been made to enhance the quality of the NMF by adding further constraints to the decomposition, such as sparsity [9], spatial localization [10, 11], and smoothness [11, 12], or by extending the model to be convolutive [13, 14]. Many extended NMF methods are derived by adding appropriate constraints and penalty terms to a cost function. Alternatively, NMF methods can be derived in a probabilistic setting, based on the distribution of the data [6, 15–17]. These approaches have the advantage that the underlying assumptions in the model are made explicit.

In this paper, we present a general method for using prior knowledge to improve the quality of the solutions in NMF. The method is derived in a probabilistic setting, and it is based on defining prior probability distributions of the factors in the NMF model in a Gaussian process framework. We assume that the nonnegative factors in the NMF are linked by a strictly increasing function to an underlying Gaussian process, specified by its covariance function. By specifying the covariance of the underlying process, we can compute NMF decompositions that agree with our prior knowledge of the factors, such as sparseness, smoothness, and symmetries. We refer to the proposed method as nonnegative matrix factorization with Gaussian process priors, or GPP-NMF for short.

#### 2. NMF with Gaussian Process Priors

In the following, we derive a method for including prior information in an NMF decomposition by assuming Gaussian process priors (for a general introduction to Gaussian processes, see, e.g., Rasmussen and Williams [18].) In our approach, the Gaussian process priors are linked to the nonnegative factors in the NMF by a suitable link function. To setup the notation, we start by deriving the standard NMF method as a maximum likelihood (ML) estimator and then move on to the maximum a posteriori (MAP) estimator. Then, we discuss Gaussian process priors and introduce a change of variable that gives better optimization properties. Finally, we discuss the selection of the link function.

##### 2.1. Maximum Likelihood NMF

The NMF problem can be stated as where is a data matrix that is factorized as the product of two element-wise nonnegative matrices, and , where denotes the nonnegative reals. The matrix is the residual noise.

There exists a number of different algorithms [8, 15–17, 19, 20, 21] for computing this factorization, some of which can be viewed as maximum likelihood methods under certain assumptions about the distribution of the data. For example, least squares NMF corresponds to i.i.d. Gaussian noise [17], and Kullback-Leibler NMF corresponds to a Poisson process [6].

The ML estimate of and is given by
where is the negative log likelihood of the
factors.
*Example 1 (least squares NMF). * An example of a maximum likelihood NMF is the
least squares estimate. If the noise is i.i.d. Gaussian with variance ,
the likelihood of the factors and can be written as
The negative log likelihood,
which serves as a cost function for optimization, is then
where we use the proportionality
symbol to denote equality subject to an additive constant. To compute a maximum likelihood estimate of and ,
the gradient of the negative log likelihood is useful:
and the gradient with respect to ,
which is easy to derive, is similar because of the symmetry of the NMF problem.

The ML estimate can be computed by multiplicative update rules based on the gradient [8], projected gradient descent [19], alternating least squares [20], Newton-type methods [21], or any other appropriate constrained optimization method.

##### 2.2. Maximum a Posteriori NMF

In this paper,
we propose a method to build prior knowledge into the solution of the NMF
problem. We choose a prior distribution over the factors in the model, that captures
our prior beliefs and uncertainties of the solution we seek. We then compute
the maximum a posteriori (MAP) estimate of the factors. Using Bayes' rule, the
posterior is given by
Since the numerator is constant,
the negative log posterior is the sum of a likelihood term that penalizes model
misfit, and a prior term that penalizes solutions that are unlikely under the
prior:
The MAP estimate of and is
and it can again be computed
using any appropriate optimization algorithm.
*Example 2 (nonnegative sparse coding). *An example
of a MAP NMF is nonnegative sparse coding (NNSC) [9, 22], where the prior on is i.i.d. exponential, and the prior on is flat with each column constrained to have
unit norm
where is the Euclidean norm of the column of .
This corresponds to the following penalty term in the cost
function:
The gradient of the negative
log prior with respect to is then
and the gradient with respect to is zero, with the further normalization
constraint given in (9).

##### 2.3. Gaussian Process Priors

In the following, we derive the MAP estimate under the assumption that the nonnegative matrices and are independently determined by a Gaussian process [18] connected by a link function. The Gaussian process framework provides a principled and practical approach to the specification of the prior probability distribution of the factors in the NMF model. The prior is specified in terms of two functions: (i) a covariance function that describes corellations in the factors and (ii) a link function, that transforms the Gaussian process prior into a desired distribution over the nonnegative reals.

We assume that and are independent, so that we may write In the following, we consider only the prior for , since the treatment of is equivalent due to the symmetry of the NMF problem. We assume that there is an underlying variable vector, , which is zero-mean multivariate Gaussian with covariance matrix : and linked to via a link function, : as which operates element-wise on its input. The function in the expression stacks its matrix operand column by column. The link function should be strictly increasing, which ensures that the inverse exists. Later, we will further assume that the derivatives of and exist. Under these assumptions, the prior over is given by (using the change of variables theorem) where denotes the Jacobian determinant and is the derivative of the link function. The negative log prior is then This expression can be combined with an appropriate likelihood function, such as the least-squares likelihood in (4) and can be optimized to yield the MAP solution; however, in our experiments, we found that a more simple and robust algorithm can be obtained by making a change of variable as explained next.

##### 2.4. Change of Optimization Variable

Instead of optimizing over the nonnegative factors and , we introduce the variables and , which are related to and by where the function maps its vector input into a matrix of appropriate size. The matrices and are the matrix square roots (Cholesky decompositions) of the covariance matrices and , such that and are standard i.i.d. Gaussian.

This change of variable serves two purposes. First of all, we found that optimizing over the transformed variables was faster, more robust, and less prone to getting stuck in local minima. Second, the transformed variables are not constrained to be nonnegative, which allows us to use existing unconstrained optimization methods to compute their MAP estimate.

The prior distribution of the transformed variable is
and the negative log prior
is
To compute the MAP estimate of
the transformed variables, we must combine this expression for the prior (and a
similar expression for the prior of ) with a likelihood function, in which the
same change of variable is made
Then, the MAP solution can be
found by optimizing over and as
and, finally, estimates of and can be computed using (17).
*Example 3 (least squares nonnegative matrix factorization with Gaussian process priors (GPP-NMF)). *If we use the least squares
likelihood in (4), the posterior distribution in (20) is given
by
The MAP estimate of and is found by minimizing this expression, for
which the derivative is useful
where denotes the Hadamard (element-wise) product.
The derivative with respect to is similar due to the symmetry of the NMF
problem.

##### 2.5. Link Function

Any strictly increasing link function that maps the nonnegative reals to the real line can be used in the proposed framework; however, in order to have a better probabilistic interpretation of the prior distribution, we propose a simple principle for choosing the link function. We choose the link function such that maps the marginal distribution of the elements of the underlying Gaussian process vector into a specifically chosen marginal distribution of the elements of . Such an inverse function can be found as , where denotes the marginal cumulative distribution functions (CDFs).

Since the marginals of a Gaussian process are
Gaussian, is the Gaussian (CDF), and, using (13), the
inverse link function is given by
where is the th diagonal element of and is the error function.
*Example 4 (exponential-to-Gaussian link function). *If we choose to have
exponential marginals in ,
as in NNSC described in Example 2, we select as the exponential CDF. The inverse link
function is then
where is an inverse scale parameter. The derivative
of the inverse link function, which is needed for the parameter estimation, is
given by
*Example 5 (rectified-Gaussian-to-Gaussian link function). *Another interesting
nonnegative distribution is the rectified Gaussian given by
Using this pdf in (24), the
inverse link function is
and the derivative of the
inverse link function is

##### 2.6. Summary of the GPP-NMF Method

The GPP-NMF method can be summarized in the following steps.

(i)Choose a suitable negative log-likelihood function based on knowledge of the distribution of the data or the residual.(ii)For each of the nonnegative factors and , choose suitable link and covariance functions according to your prior beliefs. If necessary, draw samples from the prior distribution to examine its properties.(iii)Compute the MAP estimate of and by (21) using any suitable unconstrained optimization algorithm.(iv)Compute and using (17). Our Matlab implemention of the GPP-NMF method is available online [23].

#### 3. Experimental Results

We will demonstrate the proposed method on two examples, first a toy example, and second an example taken from the chemical shift brain imaging literature.

In our experiments, we use the least squares GPP-NMF described in Example 3 and the link functions described in Examples 4-5. The specific optimization method used to compute the GPP-NMF MAP estimate is not the topic of this paper, and any unconstrained optimization algorithm could in principle be used. In our experiments, we used a simple gradient descent with line search to perform a total of 1000 alternating updates of and , after which the algorithm had converged. For details of the implementation, see the accompanying Matlab code [23].

##### 3.1. Toy Example

We generated a data matrix, , by taking a random sample from the GPP-NMF model with two factors. We chose the generating covariance function for both and as a Gaussian radial basis function (RBF) where and are two sample indices, and the length-scale parameter, which determines the smoothness of the factors, was . We set the covariance between the two factors to zero, such that the factors were uncorrelated. For the matrix we used the rectified-Gaussian-to-Gaussian link function with ; and for we used the exponential-to-Gaussian link function with . Finally, we added independent Gaussian noise with variance , which resulted in a signal-to-noise ratio of approximately dB. The generated data matrix is shown in Figure 1.

We then decomposed the generated data matrix using the following four different methods:

(i)** LS-NMF:** standard least squares NMF
[8]. This algorithm
does not allow negative data points, so these were set to zero in the
experiment.(ii)

**constrained NMF [6, 7], which is a least squares NMF algorithm that allows negative observations.(iii)**

*CNMF*:**the proposed method with correct link functions, covariance matrix, and parameter values.(iv)**

*GPP-NMF*:*correct prior*:**to illustrate the sensitivity of the method to prior assumptions, we evaluated the proposed method with incorrect prior assumptions: we switched the link functions, such that for we used the exponential-to-Gaussian, and for we used the rectified-Gaussian-to-Gaussian. We used an RBF covariance function with and for and respectively.**

*GPP-NMF*:*incorrect prior*:The results of the decompositions are shown as reconstructed data matrices in Figure 1. All four methods find solutions that visually appear to fit the underlying data. Both LS-NMF and CNMF find nonsmooth solutions, whereas the two GPP-NMF results are smooth in accordance with the priors. In the GPP-NMF with incorrect prior, the dark areas (high-pixel intensities) appear too wide in the first axis direction and too narrow in the section axis direction, due to the incorrect setting of the covariance function. The GPP-NMF with correct prior is visually almost equal to the true underlying data.

Plots of the estimated factors are show in Figure 2. The factors estimated by the LS-NMF and the CNMF methods appear noisy and are nonsmooth, whereas the factors estimated by the GPP-NMF are smooth. The factors estimated by the LS-NMF method have a positive bias, because of the truncation of negative data. The GPP-NMF with incorrect prior has too many local extrema in the factor and too few in the factor due to the incorrect covariance functions. There are only minor difference between the result of the GPP-NMF with the correct prior and the underlying factors.

Measures of root mean squared error (RMSE) of the four decompositions are given in Figure 3. All four methods fit the noisy data almost equally well. (Note that, due to the additive noise with variance 25, a perfect fit to the underlying factors would result in an RMSE of with respect to the noisy data.) The LS-NMF fits the data worst due to the truncation of negative data points, and the CNMF fits the data best, due to overfitting. With respect to the noise-free data and the underlying factors, the RMSE is worst for the LS-NMF and best for the GPP-NMF with correct prior. The GPP-NMF with incorrect prior is better than both LS-NMF and CNMF in this case. This shows that in this situation it is better to use a prior which is not perfectly correct, compared to using no prior as in the LS-NMF and CNMF methods, (which corresponds to a flat prior over the nonnegative reals and no correlations).

##### 3.2. Chemical Shift Srain Imaging Example

Next, we
demonstrate the GPP-NMF method on ^{1}H decoupled ^{31}P chemical shift imaging data of the human brain. We use the data set
from Ochs et al. [24],
which has also been analyzed by Sajda et al. [6, 7]. The data set, which is
shown in Figure 4, consists of spectra measured on an grid in the brain.

Ochs et al. [24] use PCA to determine that the data set is adequately described by two sources (which correspond to brain and muscle tissue). They propose a bilinear Bayesian approach, in which they use a smooth prior over the constituent spectra, and force to zero the amplitude of the spectral shape corresponding to muscle tissue at 12 positions deep inside the head. Their approach produces physically plausible results, but it is computationally very expensive and takes several hours to compute.

Sajda et al. [6, 7] propose an NMF approach that is reported also to produce physically plausible results. Their method is several orders of magnitude faster, taking less than a second to compute. The disadvantage of the method of Sajda et al. compared to the Bayesian approach of Ochs et al. is that it provides no mechanism for using prior knowledge to improve the solution.

The GPP-NMF approach we propose in this paper bridges the gap between the two previous approaches, in the sense that it is a relatively fast NMF approach, in which priors over the factors can be specified. These priors are specified by the choice of the link and covariance functions. We used prior predictive sampling to find reasonable settings of the the function parameters: we drew random samples from the prior distribution and examined properties of the factors and reconstructed data. We then manually adjusted the parameters of the prior to match our prior beliefs. An example of a random draw from the prior distribution is shown in Figure 5, with the parameters set as described below.

We assumed that the factors are uncorrelated, so the covariance between factors is zero. We used a Gaussian RBF covariance function for the constituent spectra, with a length scale of parts per million (ppm), and we used the exponential-to-Gaussian link function with . This gave a prior for the spectra that is sparse with narrow smooth peaks. For the amplitude at the 512 voxels in the head, we used a Gaussian RBF covariance function on the 3D voxel indices, with length scale . Furthermore, we centered the left-to-right coordinate axis in the middle of the brain, and computed the RBF kernel on the absolute value of this coordinate, so that a left-to-right symmetry was introduced in the prior distribution. Again, we used the exponential-to-Gaussian link function, and we chose to match the overall magnitude of the data. This gave a prior for the amplitude distribution that is sparse, smooth, and symmetric. The noise variance was set to which corresponds to the noise level in the data set.

We then decomposed the data set using the proposed GPP-NMF algorithm and, for comparison, reproduced the results of Sajda et al. [7] using their CNMF method. The results of the experiments are shown in Figure 4. An example of a random draw from the prior distribution is shown in Figure 5. The results of the CNMF is shown in Figure 6, and the results of the GPP-NMF is shown in Figure 7. The figures show the constituent spectra and the fifth axial slice of the spatial distribution of the spectra. The spatial distributions are smoothed in the illustration, similar to the way the results are visualized in the literature.

The results show that both methods give physically plausible results. The main difference is that the spatial distribution of the spectra corresponding to muscle and brain tissue is much more separated in the GPP-NMF result, which is due to the exponential, smooth, and symmetric prior distribution. By including prior information, we obtain a solution, where the factor corresponding to muscle tissue is clearly located on the edge of the skull.

#### 4. Conclusions

We have introduced a general method for exploiting prior knowledge in nonnegative matrix factorization, based on Gaussian process priors, linked to the nonnegative factors by a link function. The method can be combined with any existing NMF cost function that has a probabilistic interpretation, and any existing unconstrained optimization algorithm can be used to compute the maximum a posteriori estimate.

Experiments on toy data show that with a suitable selection of the prior distribution of the nonnegative factors, the GPP-NMF method gives much better results in terms of estimating the true underlying factors, both when compared to traditional NMF and CNMF.

Experiments on chemical shift brain imaging data show that the GPP-NMF method can be successfully used to include prior knowledge of the spectral and spatial distribution, resulting in better spatial separation between spectra corresponding to muscle and brain tissue.

#### Acknowledgments

We would like to thank Paul Sajda and Truman Brown for making the brain imaging data available to us. This research was supported by the Intelligent Sound project, Danish Technical Research Council Grant no. 26–02–0092 and partly supported also by the European Commission through the sixth framework IST Network of Excellence: Pattern Analysis, Statistical Modelling and Computational Learning (PASCAL), Contract no. 506778.

#### References

- P. Paatero and U. Tapper, “Positive matrix factorization: a nonnegative factor model with optimal utilization of error estimates of data values,”
*Environmetrics*, vol. 5, no. 2, pp. 111–126, 1994. View at Publisher · View at Google Scholar - D. D. Lee and H. S. Seung, “Learning the parts of objects by nonnegative matrix factorization,”
*Nature*, vol. 401, no. 6755, pp. 788–791, 1999. View at Publisher · View at Google Scholar - L. Weixiang, Z. Nanning, and Y. Qubo, “Nonnegative matrix factorization and its applications in pattern recognition,”
*Chinese Science Bulletin*, vol. 51, no. 1, pp. 7–18, 2006. View at Publisher · View at Google Scholar - C. Ding, X. He, and H. D. Simon, “On the equivalence of nonnegative matrix factorization and spectral clustering,” in
*Proceedings of the 5th SIAM International Conference on Data Mining*, pp. 606–610, Newport Beach, Calif, USA, April 2005. - S. Tsuge, M. Shishibori, S. Kuroiwa, and K. Kita, “Dimensionality reduction using nonnegative matrix factorization for information retrieval,” in
*Proceedings of IEEE International Conference on Systems, Man and Cybernetics*, vol. 2, pp. 960–965, Tucson, Ariz, USA, October 2001. - P. Sajda, S. Du, and L. C. Parra, “Recovery of constituent spectra using nonnegative matrix factorization,” in
*Wavelets: Applications in Signal and Image Processing*, vol. 5207 of*Proceedings of SPIE*, pp. 321–331, San Diego, Calif, USA, August 2003. View at Publisher · View at Google Scholar - P. Sajda, S. Du, T. R. Brown et al., “Nonnegative matrix factorization for rapid recovery of constituent spectra in magnetic resonance chemical shift imaging of the brain,”
*IEEE Transactions on Medical Imaging*, vol. 23, no. 12, pp. 1453–1465, 2004. View at Publisher · View at Google Scholar - D. D. Lee and H. S. Seung, “Algorithms for nonnegative matrix factorization,” in
*Proceedings of the 13th Annual Conference on Neural Information Processing Systems (NIPS '00)*, pp. 556–562, Denver, Colo, USA, November-December 2000. - P. Hoyer, “Nonnegative sparse coding,” in
*Proceedings of the 12th IEEE Workshop on Neural Networks for Signal Processing (NNSP '02)*, pp. 557–565, Martigny, Switzerland, September 2002. View at Publisher · View at Google Scholar - S. Z. Li, X. W. Hou, H. J. Zhang, and Q. S. Cheng, “Learning spatially localized, parts-based representation,” in
*Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR '01)*, vol. 1, pp. 207–212, Kauai, Hawaii, USA, December 2001. View at Publisher · View at Google Scholar - Z. Chen and A. Cichocki, “Nonnegative matrix factorization with temporal smoothness and/or spatial decorrelation constraints,”
*Technical Report*, Laboratory for Advanced Brain Signal Processing, RIKEN, Tokyo, Japan, 2005. View at Google Scholar - T. Virtanen, “Sound source separation using sparse coding with temporal continuity objective,” in
*Proceedings of the International Computer Music Conference (ICMC '03)*, pp. 231–234, Singapore, September-October 2003. - P. Smaragdis, “Nonnegative matrix factor deconvolution; extraction of multiple sound sources from monophonic inputs,” in
*Proceedings of the 5th International Conference on Independent Component Analysis and Blind Signal Separation (ICA '04)*, vol. 3195 of*Lecture Notes in Computer Science*, pp. 494–499, Springer, Granada, Spain, September 2004. - M. N. Schmidt and M. Mørup, “Nonnegative matrix factor 2-d deconvolution for blind single channel source separation,” in
*Proceedings of the 6th International Conference on Independent Component Analysis and Blind Signal Separation (ICA '06)*, vol. 3889 of*Lecture Notes in Computer Science*, pp. 700–707, Springer, Charleston, SC, USA, March 2006. View at Publisher · View at Google Scholar - O. Winther and K. B. Petersen, “Bayesian independent component analysis: variational methods and nonnegative decompositions,”
*Digital Signal Processing*, vol. 17, no. 5, pp. 858–872, 2007. View at Publisher · View at Google Scholar - T. Hofmann, “Probabilistic latent semantic indexing,” in
*Proceedings of the 22nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR '99)*, pp. 50–57, Berkeley, Calif, USA, August 1999. View at Publisher · View at Google Scholar - A. Cichocki, R. Zdunek, and S.-I. Amari, “Csiszar's divergences for nonnegative matrix factorization: family of new algorithms,” in
*Proceedings of the 6th International Conference on Independent Component Analysis and Blind Signal Separation (ICA '06)*, vol. 3889 of*Lecture Notes in Computer Science*, pp. 32–39, Springer, Charleston, SC, USA, March 2006. View at Publisher · View at Google Scholar - C. E. Rasmussen and C. K. I. Williams,
*Gaussian Processes for Machine Learning*, MIT Press, Cambridge, Mass, USA, 2006. - C.-J. Lin, “Projected gradient methods for nonnegative matrix factorization,”
*Neural Computation*, vol. 19, no. 10, pp. 2756–2779, 2007. View at Publisher · View at Google Scholar - M. W. Berry, M. Browne, A. N. Langville, V. P. Pauca, and R. J. Plemmons, “Algorithms and applications for approximate nonnegative matrix factorization,”
*Computational Statistics & Data Analysis*, vol. 52, no. 1, pp. 155–173, 2007. View at Publisher · View at Google Scholar - D. Kim, S. Sra, and I. S. Dhillon, “Fast Newton-type methods for the least squares nonnegative matrix approximation problem,” in
*Proceedings of the 7th SIAM International Conference on Data Mining*, pp. 343–354, Minneapolis, Minn, USA, April 2007. - J. Eggert and E. Körner, “Sparse coding and NMF,” in
*Proceedings of IEEE International Joint Conference on Neural Networks (IJCNN '04)*, vol. 4, pp. 2529–2533, Budapest, Hungary, July 2004. View at Publisher · View at Google Scholar - M. N. Schmidt and H. Laurberg, “Nonnegative matrix factorization with Gaussian process priors,”
*Computational Intelligence and Neuroscience*. In press. View at Publisher · View at Google Scholar - M. F. Ochs, R. S. Stoyanova, F. Arias-Mendoza, and T. R. Brown, “A new method for spectral decomposition using a bilinear Bayesian approach,”
*Journal of Magnetic Resonance*, vol. 137, no. 1, pp. 161–176, 1999. View at Google Scholar