Compressive Mahalanobis Metric Learning Adapts to Intrinsic Dimension
arxiv(2023)
摘要
Metric learning aims at finding a suitable distance metric over the input
space, to improve the performance of distance-based learning algorithms. In
high-dimensional settings, it can also serve as dimensionality reduction by
imposing a low-rank restriction to the learnt metric. In this paper, we
consider the problem of learning a Mahalanobis metric, and instead of training
a low-rank metric on high-dimensional data, we use a randomly compressed
version of the data to train a full-rank metric in this reduced feature space.
We give theoretical guarantees on the error for Mahalanobis metric learning,
which depend on the stable dimension of the data support, but not on the
ambient dimension. Our bounds make no assumptions aside from i.i.d. data
sampling from a bounded support, and automatically tighten when benign
geometrical structures are present. An important ingredient is an extension of
Gordon's theorem, which may be of independent interest. We also corroborate our
findings by numerical experiments.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要