Random projection below the JL limit
The Johnson-Lindenstrauss (JL) lemma, with known probability, sets a lower bound q 0 on the dimension for which a random projection of p-dimensional vector data is guaranteed to be within (1±ε) of being an isometry in a randomly projected downspace. We study several ways to identify a "good&quo...
Saved in:
Published in | 2016 International Joint Conference on Neural Networks (IJCNN) pp. 2414 - 2423 |
---|---|
Main Authors | , , , , |
Format | Conference Proceeding |
Language | English |
Published |
IEEE
01.07.2016
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | The Johnson-Lindenstrauss (JL) lemma, with known probability, sets a lower bound q 0 on the dimension for which a random projection of p-dimensional vector data is guaranteed to be within (1±ε) of being an isometry in a randomly projected downspace. We study several ways to identify a "good" rogue random projection when the target downspace has dimensions below the JL limit. The tools used towards this end are Pearson and Spearman correlation coefficients, and a visual imaging method (a cluster heat map) that usually reveals cluster structure in spaces of any dimension. We use four synthetic data sets and the ubiquitous Iris data to study our procedures for tracking the reliability of RRPs. Unsurprisingly, rogue random projection is quite unpredictable. At its best, it is every bit as good as Principal Components Analysis, but at it's worst, it is awful. Pearson and Spearman correlations do signal good and bad projections, but the visual imaging method seems even more effective in determining the quality of RRPs. |
---|---|
ISSN: | 2161-4407 |
DOI: | 10.1109/IJCNN.2016.7727499 |