GPU-Accelerated Computation of the Non-Negative Matrix Factorization
Recommend Documents
Suppose one wants to model moustaches and beards of a database of human faces with. 4 parameters. The first parameter represents the top of the moustache ...
ristretto, written in Python, which allows the reproduction of all results (GIT .... decomposition of Y is used to form a matrix Q â RmÃk with orthogonal ...... database of handwritten digits, which comprises 60,000 training and 10,000 ... Table 4
Nov 5, 2014 - are from Wallflower [40]. The first row image sequences are from [21]. include each image frame as a rank and thus it cause dense basis matrix ...
perimental results on real-world datasets show that our method has strong perfor- ..... other hand, some other methods can find more accurate results but require .... 11K. 0.80. 0.82. 0.80. 0.77. 0.80 0.86. 0.82. 0.52. 0.80. 0.87. PROTEIN. 18K.
However, most of the existing NMF algorithms do not pro- vide uniqueness of the ..... Second, the Welch's method is applied to calculate the .... (where â and â denote the elementwise multiplication and division, respectively), it is obvious that
Department of Information and Computer Science. â. Aalto University School of Science and Technology. P.O.Box 15400 ... Nonnegative learning based on matrix factorization has received a lot ...... Mathematical Forum, 3:1853â1870, 2008.
Sep 22, 2017 - methods to fuse hyperspectral data and high spatial resolution .... Remote Sens. , vol. 43, pp. 455-465, 2005. N. Yokoya, C. Grohnfeldt, and J. Chanussot, "Hyperspectral and ... Geoscience and Remote Sensing Magazine, vol.
Exploratory Multi-way Data Analysis and Blind Source Separation,. Wiley, 09 ..... Proc. of the 8th IEEE Int. Conf. on Data Mining (ICDM), pp. 353-362, 2008.
ABSTRACT. We address the problem of blind audio source separation in the under-determined and convolutive case. The cont
... HUT, Finland. E-mail: {zhijian.yuan, zhirong.yang, erkki.oja}@hut.fi ... The close relation of NMF to clustering has been shown by Ding et al [4]. For recent ...
value decomposition and nonnegative matrix factorization for the purpose of ... topic that has been used. Nirmal Thapa is with Department of Computer Science, University of .... Mathematical derivation for update formula. Let,. Q = A â HW2. F.
imaging (CSI), hierarchical decomposition, magnetic resonance. (MR), magnetic ... lege of New York, New York, NY 10031 USA (e-mail: [email protected]).
Feb 10, 2010 - Representation of convolutive mixing system and formulation of Multichannel NMF problem. ...... vibrato). The EM and MU algorithms were run for 500 itera- ... guitar) is recorded with several microphones placed close to.
Abstract+An activeSperiodSaware supervised nonnegative maS trix factorization (NMF) approach for music transcription is proposed. supervised NMF relies on ...
Therefore, data-driven dictionary is introduced into sparse representation [22]. Dictionary learning ... Multi-task learning is a type of machine learning that learns.
Sep 19, 2016 - Facebook ..... OF GTNMF CO MPAR ED TO S EVE RA L NMF APP ROAC HES . ..... ternational ACM SIGIR conference on Research and development in .... [21] U. Von Luxburg, âA tutorial on spectral clustering,â Statistics and.
SSNMF can use the (partial) class label explicitly, serving as semi-supervised feature .... 1) Document Data Matrix: In the vector-space model of text data, each ... the power spectral density (PSD) in the band 8â30 Hz every. 62.5 ms, (i.e., 16 ...
Nov 3, 2011 - (MC)jSj Ti (MR)i = M ...... Say a row Mj is a loner if (ignoring other rows that are copies of Mj) it is .... For a row M j, we call it a robust-loner if upon.
and Saul, 2000] [Niyogi, 2003]. Yet NMF does not exploit ..... J. H. Friedman. ... Saul. Nonlinear dimensionality reduction by locally lin- ear embedding. Science ...
This work was supported in part by the National Science Foundation. ... Nonnegative Matrix Factorization: Algorithms and Applications ... Math, Georgia Tech.
Nonnegative Matrix Factorization (NMF) is a dimension reduction method that has been widely used for ... A mathematical formulation of NMF is given as follows.
of Non-negative Matrix Factorization (NMF) on Compute Unified Device ... tes. n x mTrain. n x r. r x mTrain. A ttrib u tes. Observations. Basis Vectors. En co d in g.
C ONTRIBUTED R ESEARCH A RTICLES
382
nmfgpu4R: GPU-Accelerated Computation of the Non-Negative Matrix Factorization (NMF) Using CUDA Capable Hardware by Sven Koitka and Christoph M. Friedrich Abstract In this work, a novel package called nmfgpu4R is presented, which offers the computation of Non-negative Matrix Factorization (NMF) on Compute Unified Device Architecture (CUDA) platforms within the R environment. Benchmarks show a remarkable speed-up in terms of time per iteration by utilizing the parallelization capabilities of modern graphics cards. Therefore the application of NMF gets more attractive for real-world sized problems because the time to compute a factorization is reduced by an order of magnitude.
Introduction Dimension reduction techniques are commonly used in machine learning and data mining tasks. For instance in text mining a corpora with thousands of words in the vocabulary could be too complex to be learned by Support Vector Machines (SVM) directly. Therefore the most important structure within the data must be extracted prior to the learning process. In the context of text mining new data axes at best represent topics in the corpora, which are used to approximate the original documents. Furthermore by reducing the feature space of the data it is less likely to be influenced by the Curse of Dimensionality (CoD) (Bellman, 1961). There are several methods to reduce the dimension of a data matrix, for example Principal Component Analysis (PCA) (Pearson, 1901) and Latent Dirichlet Allocation (LDA) (Blei et al., 2003). Another powerful technique namely Non-negative Matrix Factorization (NMF) (Lee and Seung, 1999) will be discussed in the first section of this work. Currently available NMF implementations require a prohibitively long computation time, which make the usage for real-world applications impractical. Therefore we present an implementation using the Compute Unified Device Architecture (CUDA) platform with a binding to the R environment. Furthermore the package is developed platform independent and is compatible with all three major platforms for R: Windows, Linux and Mac OS X.
Overview of non-negative matrix factorization Let X ∈ Rn+×m be a matrix with n attributes and m observations in the dataset, then the data matrix X is approximated by the product of two new matrices W and H (Lee and Seung, 2001): X ≈ WH
(1)
Each column of the matrix W ∈ Rn+×r represents a single basis vector, whereas each column of the matrix H ∈ Rr+×m represents an encoding vector. Therefore a column of the data matrix can be approximated by the linear combination of all basis vectors with one encoding vector (Lee and Seung, 2001). The importance of each basis vector can be seen by analysing the row sums of matrix H. Row sums with a low value identify basis vectors with very little influence on the dataset and vice versa (Skillicorn, 2007). It is also important to note that the data matrix as well as both matrices W and H contain only non-negative values. Besides the general convention in the context of data mining, NMF expects columns to represent observations of the dataset instead of attributes (Skillicorn, 2007), as visualized in Figure 1. For that reason it is very important to read the data matrix definition in the literature carefully. Contrary to PCA or Singular Value Decomposition (SVD), the basis vectors are not linearly independent and thus the solution is not unique. However the reconstruction of the data matrix is purely additive and yields a more natural parts-based decomposition (Lee and Seung, 1999). As the factorization should represent a compressed form of the original data matrix, one approach is to choose r depending on the number of rows and columns of the data matrix (Lee and Seung, 2001): r