Hilbert Curve Projection Distance for Distribution Comparison

Tao Li, Cheng Meng, Hongteng Xu*, Jun Yu

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

摘要

Distribution comparison plays a central role in many machine learning tasks like data classification and generative mod- eling. In this study, we propose a novel metric, called Hilbert curve projection (HCP) distance, to measure the distance between two probability distributions with low complexity. In particular, we first project two high-dimensional probability distributions using Hilbert curve to obtain a coupling between them, and then cal- culate the transport distance between these two distributions in the original space, according to the coupling. We show that HCP distance is a proper metric and is well-defined for probability measures with bounded supports. Furthermore, we demonstrate that the modified empirical HCP distance with the Lp cost in the d-dimensional space converges to its population counterpart at a rate of no more than O(n−1/2 max{d,p}). To suppress the curse-of-dimensionality, we also develop two variants of the HCP distance using (learnable) subspace projections. Experiments on both synthetic and real-world data show that our HCP distance works as an effective surrogate of the Wasserstein distance with low complexity and overcomes the drawbacks of the sliced Wasserstein distance.

源语言英语
页(从-至)4993-5007
页数15
期刊IEEE Transactions on Pattern Analysis and Machine Intelligence
46
7
DOI
出版状态已出版 - 1 7月 2024

指纹

探究 'Hilbert Curve Projection Distance for Distribution Comparison' 的科研主题。它们共同构成独一无二的指纹。

引用此