Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
[PDF][PDF] Trustllm: Trustworthiness in large language models
Large language models (LLMs), exemplified by ChatGPT, have gained considerable
attention for their excellent natural language processing capabilities. Nonetheless, these …
attention for their excellent natural language processing capabilities. Nonetheless, these …
Pandalm: An automatic evaluation benchmark for llm instruction tuning optimization
Instruction tuning large language models (LLMs) remains a challenging task, owing to the
complexity of hyperparameter selection and the difficulty involved in evaluating the tuned …
complexity of hyperparameter selection and the difficulty involved in evaluating the tuned …
[HTML][HTML] Position: TrustLLM: Trustworthiness in large language models
Large language models (LLMs) have gained considerable attention for their excellent
natural language processing capabilities. Nonetheless, these LLMs present many …
natural language processing capabilities. Nonetheless, these LLMs present many …
Logiqa 2.0—an improved dataset for logical reasoning in natural language understanding
NLP research on logical reasoning regains momentum with the recent releases of a handful
of datasets, notably LogiQA and Reclor. Logical reasoning is exploited in many probing …
of datasets, notably LogiQA and Reclor. Logical reasoning is exploited in many probing …
Glue-x: Evaluating natural language understanding models from an out-of-distribution generalization perspective
Pre-trained language models (PLMs) are known to improve the generalization performance
of natural language understanding models by leveraging large amounts of data during the …
of natural language understanding models by leveraging large amounts of data during the …
Coco-counterfactuals: Automatically constructed counterfactual examples for image-text pairs
Counterfactual examples have proven to be valuable in the field of natural language
processing (NLP) for both evaluating and improving the robustness of language models to …
processing (NLP) for both evaluating and improving the robustness of language models to …
SenticNet
SenticNet is the knowledge base which the sentic computing framework leverages on for
concept-level sentiment analysis. This chapter illustrates how such a resource is built. In …
concept-level sentiment analysis. This chapter illustrates how such a resource is built. In …
A survey on multilingual large language models: Corpora, alignment, and bias
Y Xu, L Hu, J Zhao, Z Qiu, K XU, Y Ye, H Gu - arxiv preprint arxiv …, 2024 - arxiv.org
Based on the foundation of Large Language Models (LLMs), Multilingual LLMs (MLLMs)
have been developed to address the challenges faced in multilingual natural language …
have been developed to address the challenges faced in multilingual natural language …
Socialcounterfactuals: Probing and mitigating intersectional social biases in vision-language models with counterfactual examples
While vision-language models (VLMs) have achieved remarkable performance
improvements recently there is growing evidence that these models also posses harmful …
improvements recently there is growing evidence that these models also posses harmful …
Challenges in applying explainability methods to improve the fairness of NLP models
Motivations for methods in explainable artificial intelligence (XAI) often include detecting,
quantifying and mitigating bias, and contributing to making machine learning models fairer …
quantifying and mitigating bias, and contributing to making machine learning models fairer …