Guest Speaker: Dr. Yoonkyung Lee
A Statistical View of Ranking: Midway between Classification and Regression
This talk examines the theoretical relation between loss criteria used in ranking and the optimal ranking functions driven by the criteria. We investigate the relation between AUC maximization and minimization of ranking risk under a convex loss in bipartite ranking, and characterize general conditions for ranking-calibration akin to classification-calibration. The best ranking functions under convex ranking-calibrated loss criteria are shown to produce the same ordering of instances as the likelihood ratio of the positive category to the negative category over the instance space. The result illuminates the parallel between ranking and classification in general, and suggests the notion of consistency in bipartite ranking.
The optimality of ranking algorithms is further considered in multipartite ranking through minimization of the theoretical risk which combines pairwise ranking errors of ordinal categories with differential ranking costs. The extension shows that for a certain class of convex loss functions, the optimal ranking function can be represented as a ratio of weighted conditional probability of the upper categories to the lower categories, where the weights are given by the pairwise ranking costs. The theoretical findings are illustrated with numerical examples.
This is joint work with Kazuki Uematsu.
Yoonkyung Lee is an Associate Professor in the Department of Statistics at Ohio State University. She received her PhD from the University of Wisconsin at Madison, and joined Ohio State in 2002. Her research areas are statistical learning and multivariate analysis with a focus on classification, ranking, dimensionality reduction, and kernel methods. She is generally interested in problems at the intersection of statistics and machine learning. She is an elected Fellow of the American Statistical Association.