Publication | Closed Access
Algorithms for Nonnegative Matrix Factorization with the β-Divergence
626
Citations
54
References
2011
Year
Search OptimizationLow-rank ApproximationEngineeringMachine LearningMatrix FactorizationNonnegative Matrix FactorizationComputational ComplexityComputer ScienceConvex NmfSurrogate Auxiliary FunctionLinear Optimization
The β‑divergence is a family of cost functions parameterized by a single shape parameter β that includes the Euclidean distance, Kullback‑Leibler divergence, and Itakura‑Saito divergence as special cases (β = 2, 1, 0). This letter proposes algorithms for nonnegative matrix factorization using the β‑divergence, based on a surrogate auxiliary function that majorizes the criterion. The authors develop a majorization–minimization scheme yielding β‑dependent multiplicative updates, prove its monotonicity for β∈(0,1), introduce a majorization‑equalization algorithm that takes larger steps along constant auxiliary‑function level sets, and extend the framework to penalized and convex NMF variants. Simulations on synthetic and real data show that the ME algorithm converges faster than the MM scheme.
This letter describes algorithms for nonnegative matrix factorization (NMF) with the β-divergence (β-NMF). The β-divergence is a family of cost functions parameterized by a single shape parameter β that takes the Euclidean distance, the Kullback-Leibler divergence, and the Itakura-Saito divergence as special cases (β = 2, 1, 0 respectively). The proposed algorithms are based on a surrogate auxiliary function (a local majorization of the criterion function). We first describe a majorization-minimization algorithm that leads to multiplicative updates, which differ from standard heuristic multiplicative updates by a β-dependent power exponent. The monotonicity of the heuristic algorithm can, however, be proven for β ∈ (0, 1) using the proposed auxiliary function. Then we introduce the concept of the majorization-equalization (ME) algorithm, which produces updates that move along constant level sets of the auxiliary function and lead to larger steps than MM. Simulations on synthetic and real data illustrate the faster convergence of the ME approach. The letter also describes how the proposed algorithms can be adapted to two common variants of NMF: penalized NMF (when a penalty function of the factors is added to the criterion function) and convex NMF (when the dictionary is assumed to belong to a known subspace).
| Year | Citations | |
|---|---|---|
Page 1
Page 1