Trustworthy LLMs: A survey and guideline for evaluating large language models' alignment

Y Liu, Y Yao, JF Ton, X Zhang, RGH Cheng… - ar** the model prediction
H Wei, H Zhuang, R **e, L Feng… - International …, 2023 - proceedings.mlr.press
In the presence of noisy labels, designing robust loss functions is critical for securing the
generalization performance of deep neural networks. Cross Entropy (CE) loss has been …

Weak proxies are sufficient and preferable for fairness with missing sensitive attributes

Z Zhu, Y Yao, J Sun, H Li, Y Liu - … Conference on Machine …, 2023 - proceedings.mlr.press
Evaluating fairness can be challenging in practice because the sensitive attributes of data
are often inaccessible due to privacy constraints. The go-to approach that the industry …

Like draws to like: A Multi-granularity Ball-Intra Fusion approach for fault diagnosis models to resists misleading by noisy labels

F Dunkin, X Li, C Hu, G Wu, H Li, X Lu… - Advanced Engineering …, 2024 - Elsevier
Although data-driven fault diagnosis methods have achieved remarkable results, these
achievements often rely on high-quality datasets without noisy labels, which can mislead the …

Transferring annotator-and instance-dependent transition matrix for learning from crowds

S Li, X **a, J Deng, S Gey, T Liu - IEEE Transactions on Pattern …, 2024 - ieeexplore.ieee.org
Learning from crowds describes that the annotations of training data are obtained with
crowd-sourcing services. Multiple annotators each complete their own small part of the …

Mitigating memorization of noisy labels via regularization between representations

H Cheng, Z Zhu, X Sun, Y Liu - arxiv preprint arxiv:2110.09022, 2021 - arxiv.org
Designing robust loss functions is popular in learning with noisy labels while existing
designs did not explicitly consider the overfitting property of deep neural networks (DNNs) …

Measuring and reducing llm hallucination without gold-standard answers via expertise-weighting

J Wei, Y Yao, JF Ton, H Guo, A Estornell… - arxiv preprint arxiv …, 2024 - arxiv.org
LLM hallucination, ie generating factually incorrect yet seemingly convincing answers, is
currently a major threat to the trustworthiness and reliability of LLMs. The first step towards …

FedFixer: Mitigating Heterogeneous Label Noise in Federated Learning

X Ji, Z Zhu, W **, O Gadyatskaya, Z Song… - Proceedings of the …, 2024 - ojs.aaai.org
Federated Learning (FL) heavily depends on label quality for its performance. However, the
label distribution among individual clients is always both noisy and heterogeneous. The …

Visual Objectification in Films: Towards a New AI Task for Video Interpretation

J Tores, L Sassatelli, HY Wu… - Proceedings of the …, 2024 - openaccess.thecvf.com
In film gender studies the concept of" male gaze" refers to the way the characters are
portrayed on-screen as objects of desire rather than subjects. In this article we introduce a …