There has been significant recent interest in sparse metric learning (SML) in which we simultaneously learn both a good distance metric and a low-dimensional representation. Unfortunately, the performance of existing sparse metric learning approaches is usually limited because the authors assumed certain problem relaxations or they target the SML objective indirectly. In this paper, we propose a Generalized Sparse Metric Learning method (GSML). This novel framework offers a unified view for understanding many of the popular sparse metric learning algorithms including the Sparse Metric Learning framework proposed in , the Large Margin Nearest Neighbor (LMNN) , and the D-ranking Vector Machine (D-ranking VM) . Moreover, GSML also establishes a close relationship with the Pairwise Support Vector Machine . Furthermore, the proposed framework is capable of extending many current non-sparse metric learning models such as Relevant Vector Machine (RCA)  and a state-of-the-art method proposed in  into their sparse versions. We present the detailed framework, provide theoretical justifications, build various connections with other models, and propose a practical iterative optimization method, making the framework both theoretically important and practically scalable for medium or large datasets. A series of experiments show that the proposed approach can outperform previous methods in terms of both test accuracy and dimension reduction, on six realworld benchmark datasets.
|Translated title of the contribution||GSML: A Unified Framework for Sparse Metric Learning|
|Title of host publication||International Conference on Data Mining|
|Publication status||Published - 2009|
Bibliographical noteName and Venue of Event: ICDM Miami, USA, 2009
Conference Proceedings/Title of Journal: Proceedings IEEE International Conference on Data Mining, ICDM 2009