[PDF][PDF] Trustllm: Trustworthiness in large language models

L Sun, Y Huang, H Wang, S Wu, Q Zhang… - arxiv preprint arxiv …, 2024 - mosis.eecs.utk.edu
Large language models (LLMs), exemplified by ChatGPT, have gained considerable
attention for their excellent natural language processing capabilities. Nonetheless, these …

Pandalm: An automatic evaluation benchmark for llm instruction tuning optimization

Y Wang, Z Yu, Z Zeng, L Yang, C Wang, H Chen… - arxiv preprint arxiv …, 2023 - arxiv.org
Instruction tuning large language models (LLMs) remains a challenging task, owing to the
complexity of hyperparameter selection and the difficulty involved in evaluating the tuned …

[HTML][HTML] Position: TrustLLM: Trustworthiness in large language models

Y Huang, L Sun, H Wang, S Wu… - International …, 2024 - proceedings.mlr.press
Large language models (LLMs) have gained considerable attention for their excellent
natural language processing capabilities. Nonetheless, these LLMs present many …

Logiqa 2.0—an improved dataset for logical reasoning in natural language understanding

H Liu, J Liu, L Cui, Z Teng, N Duan… - … on Audio, Speech …, 2023 - ieeexplore.ieee.org
NLP research on logical reasoning regains momentum with the recent releases of a handful
of datasets, notably LogiQA and Reclor. Logical reasoning is exploited in many probing …

Glue-x: Evaluating natural language understanding models from an out-of-distribution generalization perspective

L Yang, S Zhang, L Qin, Y Li, Y Wang, H Liu… - arxiv preprint arxiv …, 2022 - arxiv.org
Pre-trained language models (PLMs) are known to improve the generalization performance
of natural language understanding models by leveraging large amounts of data during the …

Coco-counterfactuals: Automatically constructed counterfactual examples for image-text pairs

T Le, V Lal, P Howard - Advances in Neural Information …, 2023 - proceedings.neurips.cc
Counterfactual examples have proven to be valuable in the field of natural language
processing (NLP) for both evaluating and improving the robustness of language models to …

SenticNet

E Cambria, A Hussain, E Cambria… - Sentic computing: a …, 2015 - Springer
SenticNet is the knowledge base which the sentic computing framework leverages on for
concept-level sentiment analysis. This chapter illustrates how such a resource is built. In …

A survey on multilingual large language models: Corpora, alignment, and bias

Y Xu, L Hu, J Zhao, Z Qiu, K XU, Y Ye, H Gu - arxiv preprint arxiv …, 2024 - arxiv.org
Based on the foundation of Large Language Models (LLMs), Multilingual LLMs (MLLMs)
have been developed to address the challenges faced in multilingual natural language …

Socialcounterfactuals: Probing and mitigating intersectional social biases in vision-language models with counterfactual examples

P Howard, A Madasu, T Le… - Proceedings of the …, 2024 - openaccess.thecvf.com
While vision-language models (VLMs) have achieved remarkable performance
improvements recently there is growing evidence that these models also posses harmful …

Challenges in applying explainability methods to improve the fairness of NLP models

E Balkir, S Kiritchenko, I Nejadgholi… - arxiv preprint arxiv …, 2022 - arxiv.org
Motivations for methods in explainable artificial intelligence (XAI) often include detecting,
quantifying and mitigating bias, and contributing to making machine learning models fairer …