Survey on multi-output learning

D Xu, Y Shi, IW Tsang, YS Ong… - IEEE transactions on …, 2019 - ieeexplore.ieee.org
The aim of multi-output learning is to simultaneously predict multiple outputs given an input.
It is an important learning problem for decision-making since making decisions in the real …

Cross-entropy loss functions: Theoretical analysis and applications

A Mao, M Mohri, Y Zhong - International conference on …, 2023 - proceedings.mlr.press
Cross-entropy is a widely used loss function in applications. It coincides with the logistic loss
applied to the outputs of a neural network, when the softmax is used. But, what guarantees …

Why relu networks yield high-confidence predictions far away from the training data and how to mitigate the problem

M Hein, M Andriushchenko… - Proceedings of the …, 2019 - openaccess.thecvf.com
Classifiers used in the wild, in particular for safety-critical systems, should not only have
good generalization properties but also should know when they don't know, in particular …

Spectrally-normalized margin bounds for neural networks

PL Bartlett, DJ Foster… - Advances in neural …, 2017 - proceedings.neurips.cc
This paper presents a margin-based multiclass generalization bound for neural networks
that scales with their margin-normalized" spectral complexity": their Lipschitz constant …

Smart “predict, then optimize”

AN Elmachtoub, P Grigas - Management Science, 2022 - pubsonline.informs.org
Many real-world analytics problems involve two significant challenges: prediction and
optimization. Because of the typically complex nature of each challenge, the standard …

Robust loss functions under label noise for deep neural networks

A Ghosh, H Kumar, PS Sastry - Proceedings of the AAAI conference on …, 2017 - ojs.aaai.org
In many applications of classifier learning, training data suffers from label noise. Deep
networks are learned using huge training data where the problem of noisy labels is …

Houdini: Fooling deep structured prediction models

M Cisse, Y Adi, N Neverova, J Keshet - arxiv preprint arxiv:1707.05373, 2017 - arxiv.org
Generating adversarial examples is a critical step for evaluating and improving the
robustness of learning machines. So far, most existing methods only work for classification …

Does distributionally robust supervised learning give robust classifiers?

W Hu, G Niu, I Sato… - … Conference on Machine …, 2018 - proceedings.mlr.press
Abstract Distributionally Robust Supervised Learning (DRSL) is necessary for building
reliable machine learning systems. When machine learning is deployed in the real world, its …

A theoretical analysis of NDCG type ranking measures

Y Wang, L Wang, Y Li, D He… - Conference on learning …, 2013 - proceedings.mlr.press
Ranking has been extensively studied in information retrieval, machine learning and
statistics. A central problem in ranking is to design a ranking measure for evaluation of …

Revisiting discriminative vs. generative classifiers: Theory and implications

C Zheng, G Wu, F Bao, Y Cao, C Li… - … on Machine Learning, 2023 - proceedings.mlr.press
A large-scale deep model pre-trained on massive labeled or unlabeled data transfers well to
downstream tasks. Linear evaluation freezes parameters in the pre-trained model and trains …