Variational approximation error in non-negative matrix factorization.

Journal: Neural networks : the official journal of the International Neural Network Society
Published Date:

Abstract

Non-negative matrix factorization (NMF) is a knowledge discovery method that is used in many fields. Variational inference and Gibbs sampling methods for it are also well-known. However, the variational approximation error has not been clarified yet, because NMF is not statistically regular and the prior distribution used in variational Bayesian NMF (VBNMF) has zero or divergence points. In this paper, using algebraic geometrical methods, we theoretically analyze the difference in negative log evidence (a.k.a. free energy) between VBNMF and Bayesian NMF, i.e., the Kullback-Leibler divergence between the variational posterior and the true posterior. We derive an upper bound for the learning coefficient (a.k.a. the real log canonical threshold) in Bayesian NMF. By using the upper bound, we find a lower bound for the approximation error, asymptotically. The result quantitatively shows how well the VBNMF algorithm can approximate Bayesian NMF; the lower bound depends on the hyperparameters and the true non-negative rank. A numerical experiment demonstrates the theoretical result.

Authors

  • Naoki Hayashi
    Simulation & Mining Division, NTT DATA Mathematical Systems Inc., 1F Shinanomachi Rengakan, 35, Shinanomachi, Shinjuku-ku, Tokyo, 160-0016, Japan; Department of Mathematical and Computing Science, Tokyo Institute of Technology, Mail-Box W8-42, 2-12-1, Oookayama, Meguro-ku, Tokyo, 152-8552, Japan. Electronic address: hayashi@msi.co.jp.