Robust data pruning under label noise via maximizing re-labeling accuracy
Data pruning, which aims to downsize a large training set into a small informative subset, is
crucial for reducing the enormous computational costs of modern deep learning. Though …
crucial for reducing the enormous computational costs of modern deep learning. Though …
Openmix: Exploring outlier samples for misclassification detection
Reliable confidence estimation for deep neural classifiers is a challenging yet fundamental
requirement in high-stakes applications. Unfortunately, modern deep neural networks are …
requirement in high-stakes applications. Unfortunately, modern deep neural networks are …
Meta-query-net: Resolving purity-informativeness dilemma in open-set active learning
Unlabeled data examples awaiting annotations contain open-set noise inevitably. A few
active learning studies have attempted to deal with this open-set noise for sample selection …
active learning studies have attempted to deal with this open-set noise for sample selection …
RONF: reliable outlier synthesis under noisy feature space for out-of-distribution detection
Out-of-distribution~(OOD) detection is fundamental to guaranteeing the reliability of
multimedia applications during deployment in the open world. However, due to the lack of …
multimedia applications during deployment in the open world. However, due to the lack of …
Adaptive Shortcut Debiasing for Online Continual Learning
We propose a novel framework DropTop that suppresses the shortcut bias in online
continual learning (OCL) while being adaptive to the varying degree of the shortcut bias …
continual learning (OCL) while being adaptive to the varying degree of the shortcut bias …
Examining Responsibility and Deliberation in AI Impact Statements and Ethics Reviews
The artificial intelligence research community is continuing to grapple with the ethics of its
work by encouraging researchers to discuss potential positive and negative consequences …
work by encouraging researchers to discuss potential positive and negative consequences …
Prioritizing Informative Features and Examples for Deep Learning from Noisy Data
D Park - arxiv preprint arxiv:2403.00013, 2024 - arxiv.org
In this dissertation, we propose a systemic framework that prioritizes informative features
and examples to enhance each stage of the development process. Specifically, we prioritize …
and examples to enhance each stage of the development process. Specifically, we prioritize …
Robust Calibration For Improved Weather Prediction Under Distributional Shift
In this paper, we present results on improving out-of-domain weather prediction and
uncertainty estimation as part of the\texttt {Shifts Challenge on Robustness and Uncertainty …
uncertainty estimation as part of the\texttt {Shifts Challenge on Robustness and Uncertainty …
Reduce, Reuse, and Recycle: Navigating Test-Time Adaptation with OOD-Contaminated Streams
Test-Time Adaptation (TTA) aims to quickly adapt a pre-trained Deep Neural Network (DNN)
to shifted test data from unseen distributions. Early TTA works only targeted simple and …
to shifted test data from unseen distributions. Early TTA works only targeted simple and …
[ZITATION][C] 강건하고 공정하며 확장 가능한 데이터 중심의 연속 학습
이재길, 신기정, 박찬영, 황의종, 최예지 - 정보과학회지, 2022 - dbpia.co.kr
최근 들어 딥러닝 (deep learning) 은 비약적인 발전을이루었으며, 이로 인해 이미지 분류/인식,
영상 인식, 음성 인식, 자연어 번역 등의 성능이 크게 향상되었다. 이러한 딥러닝 발전의 주요 …
영상 인식, 음성 인식, 자연어 번역 등의 성능이 크게 향상되었다. 이러한 딥러닝 발전의 주요 …