An inherently nonnegative latent factor model for high-dimensional and sparse matrices from industrial applications

Xin Luo, Mengchu Zhou, Shuai Li, Mingsheng Shang

Research output: Journal article publicationJournal articleAcademic researchpeer-review

125 Citations (Scopus)


High-dimensional and sparse (HiDS) matrices are commonly encountered in many big-data-related and industrial applications like recommender systems. When acquiring useful patterns from them, nonnegative matrix factorization (NMF) models have proven to be highly effective owing to their fine representativeness of the nonnegative data. However, current NMF techniques suffer from: 1) inefficiency in addressing HiDS matrices; and 2) constraints in their training schemes. To address these issues, this paper proposes to extract nonnegative latent factors (NLFs) from HiDS matrices via a novel inherently NLF (INLF) model. It bridges the output factors and decision variables via a single-element-dependent mapping function, thereby making the parameter training unconstrained and compatible with general training schemes on the premise of maintaining the nonnegativity constraints. Experimental results on six HiDS matrices arising from industrial applications indicate that INLF is able to acquire NLFs from them more efficiently than any existing method does.

Original languageEnglish
Pages (from-to)2011-2022
Number of pages12
JournalIEEE Transactions on Industrial Informatics
Issue number5
Publication statusPublished - 1 May 2018


  • Big data
  • high-dimensional and sparse matrix
  • learning algorithms
  • missing-data estimation
  • nonnegative latent factor analysis
  • optimization methods recommender system

ASJC Scopus subject areas

  • Control and Systems Engineering
  • Information Systems
  • Computer Science Applications
  • Electrical and Electronic Engineering

Cite this