Trustworthy llms: a survey and guideline for evaluating large language models' alignment

Y Liu, Y Yao, JF Ton, X Zhang, R Guo, H Cheng… - ar** the model prediction
H Wei, H Zhuang, R **e, L Feng… - … on machine learning, 2023 - proceedings.mlr.press
In the presence of noisy labels, designing robust loss functions is critical for securing the
generalization performance of deep neural networks. Cross Entropy (CE) loss has been …

Fedfixer: mitigating heterogeneous label noise in federated learning

X Ji, Z Zhu, W **, O Gadyatskaya, Z Song… - Proceedings of the …, 2024 - ojs.aaai.org
Federated Learning (FL) heavily depends on label quality for its performance. However, the
label distribution among individual clients is always both noisy and heterogeneous. The …

Unmasking and improving data credibility: A study with datasets for training harmless language models

Z Zhu, J Wang, H Cheng, Y Liu - arxiv preprint arxiv:2311.11202, 2023 - arxiv.org
Language models have shown promise in various tasks but can be affected by undesired
data during training, fine-tuning, or alignment. For example, if some unsafe conversations …

Imprecise label learning: A unified framework for learning with various imprecise label configurations

H Chen, A Shah, J Wang, R Tao… - Advances in …, 2025 - proceedings.neurips.cc
Learning with reduced labeling standards, such as noisy label, partial label, and
supplementary unlabeled data, which we generically refer to as imprecise label, is a …

Weak proxies are sufficient and preferable for fairness with missing sensitive attributes

Z Zhu, Y Yao, J Sun, H Li, Y Liu - … Conference on Machine …, 2023 - proceedings.mlr.press
Evaluating fairness can be challenging in practice because the sensitive attributes of data
are often inaccessible due to privacy constraints. The go-to approach that the industry …

Transferring annotator-and instance-dependent transition matrix for learning from crowds

S Li, X **a, J Deng, S Gey, T Liu - IEEE Transactions on Pattern …, 2024 - ieeexplore.ieee.org
Learning from crowds describes that the annotations of training data are obtained with
crowd-sourcing services. Multiple annotators each complete their own small part of the …

Mitigating memorization of noisy labels via regularization between representations

H Cheng, Z Zhu, X Sun, Y Liu - arxiv preprint arxiv:2110.09022, 2021 - arxiv.org
Designing robust loss functions is popular in learning with noisy labels while existing
designs did not explicitly consider the overfitting property of deep neural networks (DNNs) …