Cross-entropy loss functions: Theoretical analysis and applications
Cross-entropy is a widely used loss function in applications. It coincides with the logistic loss
applied to the outputs of a neural network, when the softmax is used. But, what guarantees …
applied to the outputs of a neural network, when the softmax is used. But, what guarantees …
Cyber security intrusion detection for agriculture 4.0: Machine learning-based solutions, datasets, and future directions
In this paper, we review and analyze intrusion detection systems for Agriculture 4.0 cyber
security. Specifically, we present cyber security threats and evaluation metrics used in the …
security. Specifically, we present cyber security threats and evaluation metrics used in the …
Designing multi-label classifiers that maximize F measures: State of the art
Multi-label classification problems usually occur in tasks related to information retrieval, like
text and image annotation, and are receiving increasing attention from the machine learning …
text and image annotation, and are receiving increasing attention from the machine learning …
-Consistency Bounds: Characterization and Extensions
A series of recent publications by Awasthi et al. have introduced the key notion of* $ H $-
consistency bounds* for surrogate loss functions. These are upper bounds on the zero-one …
consistency bounds* for surrogate loss functions. These are upper bounds on the zero-one …
Multi-Class -Consistency Bounds
We present an extensive study of $ H $-consistency bounds for multi-class classification.
These are upper bounds on the target loss estimation error of a predictor in a hypothesis set …
These are upper bounds on the target loss estimation error of a predictor in a hypothesis set …
Post-hoc estimators for learning to defer to an expert
H Narasimhan, W Jitkrittum… - Advances in …, 2022 - proceedings.neurips.cc
Many practical settings allow a learner to defer predictions to one or more costly experts. For
example, the learning to defer paradigm allows a learner to defer to a human expert, at …
example, the learning to defer paradigm allows a learner to defer to a human expert, at …
Fairness with overlap** groups; a probabilistic perspective
In algorithmically fair prediction problems, a standard goal is to ensure the equality of
fairness metrics across multiple overlap** groups simultaneously. We reconsider this …
fairness metrics across multiple overlap** groups simultaneously. We reconsider this …
Learning with complex loss functions and constraints
H Narasimhan - International Conference on Artificial …, 2018 - proceedings.mlr.press
We develop a general approach for solving constrained classification problems, where the
loss and constraints are defined in terms of a general function of the confusion matrix. We …
loss and constraints are defined in terms of a general function of the confusion matrix. We …
Consistent multilabel classification
OO Koyejo, N Natarajan… - Advances in Neural …, 2015 - proceedings.neurips.cc
Multilabel classification is rapidly develo** as an important aspect of modern predictive
modeling, motivating study of its theoretical aspects. To this end, we propose a framework …
modeling, motivating study of its theoretical aspects. To this end, we propose a framework …
Generalized test utilities for long-tail performance in extreme multi-label classification
E Schultheis, M Wydmuch… - Advances in …, 2024 - proceedings.neurips.cc
Extreme multi-label classification (XMLC) is a task of selecting a small subset of relevant
labels from a very large set of possible labels. As such, it is characterized by long-tail labels …
labels from a very large set of possible labels. As such, it is characterized by long-tail labels …