Back to Search Start Over

Orthogonal Procrustes Analysis for Dictionary Learning in Sparse Linear Representation.

Authors :
Grossi, Giuliano
Lanzarotti, Raffaella
Lin, Jianyi
Source :
PLoS ONE; 1/19/2017, Vol. 12 Issue 1, p1-16, 16p
Publication Year :
2017

Abstract

In the sparse representation model, the design of overcomplete dictionaries plays a key role for the effectiveness and applicability in different domains. Recent research has produced several dictionary learning approaches, being proven that dictionaries learnt by data examples significantly outperform structured ones, e.g. wavelet transforms. In this context, learning consists in adapting the dictionary atoms to a set of training signals in order to promote a sparse representation that minimizes the reconstruction error. Finding the best fitting dictionary remains a very difficult task, leaving the question still open. A well-established heuristic method for tackling this problem is an iterative alternating scheme, adopted for instance in the well-known K-SVD algorithm. Essentially, it consists in repeating two stages; the former promotes sparse coding of the training set and the latter adapts the dictionary to reduce the error. In this paper we present R-SVD, a new method that, while maintaining the alternating scheme, adopts the Orthogonal Procrustes analysis to update the dictionary atoms suitably arranged into groups. Comparative experiments on synthetic data prove the effectiveness of R-SVD with respect to well known dictionary learning algorithms such as K-SVD, ILS-DLA and the online method OSDL. Moreover, experiments on natural data such as ECG compression, EEG sparse representation, and image modeling confirm R-SVD’s robustness and wide applicability. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
19326203
Volume :
12
Issue :
1
Database :
Complementary Index
Journal :
PLoS ONE
Publication Type :
Academic Journal
Accession number :
120806422
Full Text :
https://doi.org/10.1371/journal.pone.0169663