Published in

Institute of Electrical and Electronics Engineers, IEEE Transactions on Image Processing, 12(24), p. 4990-5003, 2015

DOI: 10.1109/tip.2015.2478396

Links

Tools

Export citation

Search in Google Scholar

Efficient Nonnegative Tucker Decompositions: Algorithms and Uniqueness

Journal article published in 2014 by Guoxu Zhou ORCID, Andrzej Cichocki, Qibin Zhao, Shengli Xie
This paper is available in a repository.
This paper is available in a repository.

Full text: Download

Green circle
Preprint: archiving allowed
Green circle
Postprint: archiving allowed
Red circle
Published version: archiving forbidden
Data provided by SHERPA/RoMEO

Abstract

Nonnegative Tucker Decomposition (NTD) is a powerful tool to extract nonnegative parts-based and physically meaningful latent components from high-dimensional tensor data, while providing natural multiway representations. However, as the data tensor often has multiple modes and is large-scale, existing NTD algorithms suffer from very high computational complexity in terms of both storage and computation time, which has been one major obstacle for practical applications of NTD. To overcome these disadvantages, in this paper we show how low (multilinear) rank approximation (LRA) of tensors is able to significantly simplify the computation of the gradients of the cost function, upon which a family of efficient first-order NTD algorithms are developed. Besides dramatically reducing the storage complexity and running time, the new algorithms are quite flexible and robust to noise because any well-established LRA approaches can be easily applied. We also show how nonnegativity incorporating sparsity substantially improves the uniqueness property and partially alleviates the curse of dimensionality of the Tucker decompositions. Simulation results on synthetic and real-world data justify the validity and high efficiency of the proposed NTD algorithms. ; Comment: 13 pages, 8 figures