Folded-concave penalization approaches to tensor completion

Wenfei Cao, Yao Wang*, Can Yang, Xiangyu Chang, Zhi Han, Zongben Xu

*Corresponding author for this work

Research output: Contribution to journalJournal articlepeer-review

47 Citations (Scopus)

Abstract

The existing studies involving matrix or tensor completion problems are commonly under the nuclear norm penalization framework due to the computational efficiency of the resulting convex optimization problem. Folded-concave penalization methods have demonstrated surprising developments in sparse learning problems due to their nice practical and theoretical properties. To share the same light of folded-concave penalization methods, we propose a new tensor completion model via folded-concave penalty for estimating missing values in tensor data. Two typical folded-concave penalties, the minmax concave plus (MCP) penalty and the smoothly clipped absolute deviation (SCAD) penalty, are employed in the new model. To solve the resulting nonconvex optimization problem, we develop a local linear approximation augmented Lagrange multiplier (LLA-ALM) algorithm which combines a two-step LLA strategy to search a local optimum of the proposed model efficiently. Finally, we provide numerical experiments with phase transitions, synthetic data sets, real image and video data sets to exhibit the superiority of the proposed model over the nuclear norm penalization method in terms of the accuracy and robustness.

Original languageEnglish
Pages (from-to)261-273
Number of pages13
JournalNeurocomputing
Volume152
DOIs
Publication statusPublished - 25 Mar 2015

Scopus Subject Areas

  • Computer Science Applications
  • Cognitive Neuroscience
  • Artificial Intelligence

User-Defined Keywords

  • Folded-concave penalization
  • Local linear approximation
  • Nuclear norm
  • Sparse learning
  • Tensor completion

Fingerprint

Dive into the research topics of 'Folded-concave penalization approaches to tensor completion'. Together they form a unique fingerprint.

Cite this