Abstract
In this paper, we study dictionary learning (DL) approach to identify the representation of low-dimensional subspaces from high-dimensional and nonnegative data. Such representation can be used to provide an affinity matrix among different subspaces for data clustering. The main contribution of this paper is to consider both nonnegativity and sparsity constraints together in DL such that data can be represented effectively by nonnegative and sparse coding coefficients and nonnegative dictionary bases. In the algorithm, we employ the proximal point technique for the resulting DL and sparsity optimization problem. We make use of coding coefficients to perform spectral clustering (SC) for data partitioning. Extensive experiments on real-world high-dimensional and nonnegative data sets, including text, microarray, and image data demonstrate that the proposed method can discover their subspace structures. Experimental results also show that our algorithm is computationally efficient and effective for obtaining high SC performance and interpreting the clustering results compared with the other testing methods.
Original language | English |
---|---|
Article number | 6497532 |
Pages (from-to) | 1188-1199 |
Number of pages | 12 |
Journal | IEEE Transactions on Neural Networks and Learning Systems |
Volume | 24 |
Issue number | 8 |
DOIs | |
Publication status | Published - Aug 2013 |
Scopus Subject Areas
- Software
- Computer Science Applications
- Computer Networks and Communications
- Artificial Intelligence
User-Defined Keywords
- Dictionary learning (DL)
- high-dimensional data
- nonnegative data
- proximal optimization
- sparsity
- spectral clustering (SC)
- subspace structure