Back to Search Start Over

Convex optimization learning of faithful Euclidean distance representations in nonlinear dimensionality reduction.

Authors :
Ding, Chao
Qi, Hou-Duo
Source :
Mathematical Programming; Jul2017, Vol. 164 Issue 1/2, p341-381, 41p
Publication Year :
2017

Abstract

Classical multidimensional scaling only works well when the noisy distances observed in a high dimensional space can be faithfully represented by Euclidean distances in a low dimensional space. Advanced models such as Maximum Variance Unfolding (MVU) and Minimum Volume Embedding (MVE) use Semi-Definite Programming (SDP) to reconstruct such faithful representations. While those SDP models are capable of producing high quality configuration numerically, they suffer two major drawbacks. One is that there exist no theoretically guaranteed bounds on the quality of the configuration. The other is that they are slow in computation when the data points are beyond moderate size. In this paper, we propose a convex optimization model of Euclidean distance matrices. We establish a non-asymptotic error bound for the random graph model with sub-Gaussian noise, and prove that our model produces a matrix estimator of high accuracy when the order of the uniform sample size is roughly the degree of freedom of a low-rank matrix up to a logarithmic factor. Our results partially explain why MVU and MVE often work well. Moreover, the convex optimization model can be efficiently solved by a recently proposed 3-block alternating direction method of multipliers. Numerical experiments show that the model can produce configurations of high quality on large data points that the SDP approach would struggle to cope with. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
00255610
Volume :
164
Issue :
1/2
Database :
Complementary Index
Journal :
Mathematical Programming
Publication Type :
Academic Journal
Accession number :
123519706
Full Text :
https://doi.org/10.1007/s10107-016-1090-7