Cross-entropy loss functions: Theoretical analysis and applications
Cross-entropy is a widely used loss function in applications. It coincides with the logistic loss
applied to the outputs of a neural network, when the softmax is used. But, what guarantees …
applied to the outputs of a neural network, when the softmax is used. But, what guarantees …
Two-stage learning to defer with multiple experts
We study a two-stage scenario for learning to defer with multiple experts, which is crucial in
practice for many applications. In this scenario, a predictor is derived in a first stage by …
practice for many applications. In this scenario, a predictor is derived in a first stage by …
Structured prediction with stronger consistency guarantees
We present an extensive study of surrogate losses for structured prediction supported by* $
H $-consistency bounds*. These are recently introduced guarantees that are more relevant …
H $-consistency bounds*. These are recently introduced guarantees that are more relevant …
Theoretically grounded loss functions and algorithms for score-based multi-class abstention
Learning with abstention is a key scenario where the learner can abstain from making a
prediction at some cost. In this paper, we analyze the score-based formulation of learning …
prediction at some cost. In this paper, we analyze the score-based formulation of learning …
-Consistency Bounds: Characterization and Extensions
A series of recent publications by Awasthi et al. have introduced the key notion of* $ H $-
consistency bounds* for surrogate loss functions. These are upper bounds on the zero-one …
consistency bounds* for surrogate loss functions. These are upper bounds on the zero-one …
Theoretically grounded loss functions and algorithms for adversarial robustness
Adversarial robustness is a critical property of classifiers in applications as they are
increasingly deployed in complex real-world systems. Yet, achieving accurate adversarial …
increasingly deployed in complex real-world systems. Yet, achieving accurate adversarial …
Predictor-rejector multi-class abstention: Theoretical analysis and algorithms
We study the key framework of learning with abstention in the multi-class classification
setting. In this setting, the learner can choose to abstain from making a prediction with some …
setting. In this setting, the learner can choose to abstain from making a prediction with some …
Multi-Class -Consistency Bounds
We present an extensive study of $ H $-consistency bounds for multi-class classification.
These are upper bounds on the target loss estimation error of a predictor in a hypothesis set …
These are upper bounds on the target loss estimation error of a predictor in a hypothesis set …
-Consistency Bounds for Pairwise Misranking Loss Surrogates
We present a detailed study of $ H $-consistency bounds for score-based ranking. These
are upper bounds on the target loss estimation error of a predictor in a hypothesis set $ H …
are upper bounds on the target loss estimation error of a predictor in a hypothesis set $ H …
Principled approaches for learning to defer with multiple experts
We present a study of surrogate losses and algorithms for the general problem of learning to
defer with multiple experts. We first introduce a new family of surrogate losses specifically …
defer with multiple experts. We first introduce a new family of surrogate losses specifically …