Random projection below the JL limit

The Johnson-Lindenstrauss (JL) lemma, with known probability, sets a lower bound q 0 on the dimension for which a random projection of p-dimensional vector data is guaranteed to be within (1±ε) of being an isometry in a randomly projected downspace. We study several ways to identify a "good&quo...

Full description

Saved in:
Bibliographic Details
Published in2016 International Joint Conference on Neural Networks (IJCNN) pp. 2414 - 2423
Main Authors Bezdek, James, Xiuyi Ye, Popescu, Mihail, Keller, James, Zare, Alina
Format Conference Proceeding
LanguageEnglish
Published IEEE 01.07.2016
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:The Johnson-Lindenstrauss (JL) lemma, with known probability, sets a lower bound q 0 on the dimension for which a random projection of p-dimensional vector data is guaranteed to be within (1±ε) of being an isometry in a randomly projected downspace. We study several ways to identify a "good" rogue random projection when the target downspace has dimensions below the JL limit. The tools used towards this end are Pearson and Spearman correlation coefficients, and a visual imaging method (a cluster heat map) that usually reveals cluster structure in spaces of any dimension. We use four synthetic data sets and the ubiquitous Iris data to study our procedures for tracking the reliability of RRPs. Unsurprisingly, rogue random projection is quite unpredictable. At its best, it is every bit as good as Principal Components Analysis, but at it's worst, it is awful. Pearson and Spearman correlations do signal good and bad projections, but the visual imaging method seems even more effective in determining the quality of RRPs.
ISSN:2161-4407
DOI:10.1109/IJCNN.2016.7727499