谷歌浏览器插件
订阅小程序
在清言上使用

Efficient dual approach to distance metric learning.

IEEE Trans. Neural Netw. Learning Syst.(2014)

引用 41|浏览44
暂无评分
摘要
Distance metric learning is of fundamental interest in machine learning because the employed distance metric can significantly affect the performance of many learning methods. Quadratic Mahalanobis metric learning is a popular approach to the problem, but typically requires solving a semidefinite programming (SDP) problem, which is computationally expensive. The worst case complexity of solving an SDP problem involving a matrix variable of size D×D with O(D) linear constraints is about O(D(6.5)) using interior-point methods, where D is the dimension of the input data. Thus, the interior-point methods only practically solve problems exhibiting less than a few thousand variables. Because the number of variables is D(D+1)/2, this implies a limit upon the size of problem that can practically be solved around a few hundred dimensions. The complexity of the popular quadratic Mahalanobis metric learning approach thus limits the size of problem to which metric learning can be applied. Here, we propose a significantly more efficient and scalable approach to the metric learning problem based on the Lagrange dual formulation of the problem. The proposed formulation is much simpler to implement, and therefore allows much larger Mahalanobis metric learning problems to be solved. The time complexity of the proposed method is roughly O(D(3)), which is significantly lower than that of the SDP approach. Experiments on a variety of data sets demonstrate that the proposed method achieves an accuracy comparable with the state of the art, but is applicable to significantly larger problems. We also show that the proposed method can be applied to solve more general Frobenius norm regularized SDP problems approximately.
更多
查看译文
关键词
mahalanobis distance,learning (artificial intelligence),mathematical programming,matrix algebra,semidefinite programming problem,distance metric learning,lagrange dual formulation,metric learning,matrix variable,general frobenius norm regularized sdp problems,semidefinite programming (sdp),quadratic mahalanobis metric learning approach,lagrange duality,machine learning,convex optimization,interior-point methods,linear constraints,learning artificial intelligence,distance,dual,metric
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要