Академический Документы
Профессиональный Документы
Культура Документы
Abstract
[KEY]NMF incorporates the non-negativity constraint and thus obtains the parts-based representation as well as enhancing the interpretability of the issue correspondingly.
1 Introduction
One of the basic concepts deeply rooted in science and engineering is that there must be something simple, compact, and elegant playing the fundamental roles under the apparent chaos and complexity. Dimensionality reduction technique is supposed to have two basic properties: 1. The dimension of the original data should be reduced. 2. The principal components, hidden concepts, prominent features, or latent variables of the data should be identified efficaciously. The negative values of both observations and latent components are physically meaningless in many kinds of real world data, such as image, spectra, and gene data. [KEY] When NMF is interpreted as a neural network learning algorithm depicting how the visible variables are generated from the hidden ones, the parts-based representation is obtained from the additive model. A positive number indicates the presence and a zero value represents the absence of some event or component. This conforms nicely to the dualistic properties of neural activity and synaptic strengths in neurophysiology: either excitatory or inhibitory without changing sign.
It is obvious that vj is the weight coefficient of the observation xj on the columns of U, the basis vectors or the latent feature vectors of X. Hence NMF decomposes each data into the linear combination of the basis vectors. In general, L can be smaller, equal or larger than M. But there are fundamental differences in the decomposition for L<M and L>M. It is a sort of sparse coding and compressed sensing with over-complete basis when L>M. [DENIFITION] As a kind of matrix factorization model, three essential questions need answering: 1. Existence, whether the nontrivial NMF solutions exist. 2. Uniqueness, under what assumptions NMF is, at least in some sense, unique 3. Effectiveness, under what assumption NMF is able to recover the right answer. [LIMITATION] The merits of NMF, parts-based representation and sparseness included, come at the price of more complexity. Besides, SVD or PCA has always a more compact spectrum than NMF. You just cannot have the best of both worlds.
4 Sparse NMF
The sparseness constraint is helpful in improving the uniqueness of the decomposition along with enforcing a local-based representation.
7 Conclusion
Interesting open issues: 6) Determination of the number L of the basis vectors. There has been no systematic mechanism for the choice of L, which brings inconvenience for practical applications. 7) Initialization. NMF is sensitive to the initial values.