arXiv Analytics

Sign in

arXiv:1902.00033 [cs.LG]AbstractReferencesReviewsResources

Compressed Diffusion

Scott Gigante, Jay S. Stanley III, Ngan Vu, David van Dijk, Kevin Moon, Guy Wolf, Smita Krishnaswamy

Published 2019-01-31Version 1

Diffusion maps are a commonly used kernel-based method for manifold learning, which can reveal intrinsic structures in data and embed them in low dimensions. However, as with most kernel methods, its implementation requires a heavy computational load, reaching up to cubic complexity in the number of data points. This limits its usability in modern data analysis. Here, we present a new approach to computing the diffusion geometry, and related embeddings, from a compressed diffusion process between data regions rather than data points. Our construction is based on an adaptation of the previously proposed measure-based (MGC) kernel that robustly captures the local geometry around data points. We use this MGC kernel to efficiently compress diffusion relations from pointwise to data region resolution. Finally, a spectral embedding of the data regions provides coordinates that are used to interpolate and approximate the pointwise diffusion map embedding of data. We analyze theoretical connections between our construction and the original diffusion geometry of diffusion maps, and demonstrate the utility of our method in analyzing big datasets, where it outperforms competing approaches.

Comments: 4 pages double column, submitted to SampTA 2019, appendix included for proofs
Categories: cs.LG, stat.ML
Related articles: Most relevant | Search more
arXiv:2106.15662 [cs.LG] (Published 2021-06-29)
Exponential Weights Algorithms for Selective Learning
arXiv:2103.08493 [cs.LG] (Published 2021-03-15)
How Many Data Points is a Prompt Worth?
arXiv:1802.03936 [cs.LG] (Published 2018-02-12)
On the Needs for Rotations in Hypercubic Quantization Hashing