Abstract
We consider the problem of rank loss minimization in the setting of multilabel classification, which is usually tackled by means of convex surrogate losses defined on pairs of labels. Very recently, this approach was put into question by a negative result showing that commonly used pairwise surrogate losses, such as exponential and logistic losses, are inconsistent. In this paper, we show a positive result which is arguably surprising in light of the previous one: the simpler univariate variants of exponential and logistic surrogates (i.e., defined on single labels) are consistent for rank loss minimization. Instead of directly proving convergence, we give a much stronger result by deriving regret bounds and convergence rates. The proposed losses suggest efficient and scalable algorithms, which are tested experimentally.
Item Type: | Conference or Workshop Item (Paper) |
---|---|
Faculties: | Mathematics, Computer Science and Statistics > Computer Science > Artificial Intelligence and Machine Learning |
Subjects: | 000 Computer science, information and general works > 004 Data processing computer science |
Annotation: | ISBN 978-1-4503-1285-1 |
Language: | English |
Item ID: | 91822 |
Date Deposited: | 06. Apr 2022, 10:12 |
Last Modified: | 17. Apr 2022, 22:31 |