Survey on multi-output learning
The aim of multi-output learning is to simultaneously predict multiple outputs given an input.
It is an important learning problem for decision-making since making decisions in the real …
It is an important learning problem for decision-making since making decisions in the real …
Cross-entropy loss functions: Theoretical analysis and applications
Cross-entropy is a widely used loss function in applications. It coincides with the logistic loss
applied to the outputs of a neural network, when the softmax is used. But, what guarantees …
applied to the outputs of a neural network, when the softmax is used. But, what guarantees …
Why relu networks yield high-confidence predictions far away from the training data and how to mitigate the problem
Classifiers used in the wild, in particular for safety-critical systems, should not only have
good generalization properties but also should know when they don't know, in particular …
good generalization properties but also should know when they don't know, in particular …
Spectrally-normalized margin bounds for neural networks
This paper presents a margin-based multiclass generalization bound for neural networks
that scales with their margin-normalized" spectral complexity": their Lipschitz constant …
that scales with their margin-normalized" spectral complexity": their Lipschitz constant …
Smart “predict, then optimize”
Many real-world analytics problems involve two significant challenges: prediction and
optimization. Because of the typically complex nature of each challenge, the standard …
optimization. Because of the typically complex nature of each challenge, the standard …
Robust loss functions under label noise for deep neural networks
In many applications of classifier learning, training data suffers from label noise. Deep
networks are learned using huge training data where the problem of noisy labels is …
networks are learned using huge training data where the problem of noisy labels is …
Houdini: Fooling deep structured prediction models
Generating adversarial examples is a critical step for evaluating and improving the
robustness of learning machines. So far, most existing methods only work for classification …
robustness of learning machines. So far, most existing methods only work for classification …
Does distributionally robust supervised learning give robust classifiers?
Abstract Distributionally Robust Supervised Learning (DRSL) is necessary for building
reliable machine learning systems. When machine learning is deployed in the real world, its …
reliable machine learning systems. When machine learning is deployed in the real world, its …
A theoretical analysis of NDCG type ranking measures
Ranking has been extensively studied in information retrieval, machine learning and
statistics. A central problem in ranking is to design a ranking measure for evaluation of …
statistics. A central problem in ranking is to design a ranking measure for evaluation of …
Revisiting discriminative vs. generative classifiers: Theory and implications
A large-scale deep model pre-trained on massive labeled or unlabeled data transfers well to
downstream tasks. Linear evaluation freezes parameters in the pre-trained model and trains …
downstream tasks. Linear evaluation freezes parameters in the pre-trained model and trains …