Prompt sapper: a LLM-empowered production tool for building AI chains

Y Cheng, J Chen, Q Huang, Z **ng, X Xu… - ACM Transactions on …, 2024 - dl.acm.org
The emergence of foundation models, such as large language models (LLMs) GPT-4 and
text-to-image models DALL-E, has opened up numerous possibilities across various …

Unifying Bias and Unfairness in Information Retrieval: A Survey of Challenges and Opportunities with Large Language Models

S Dai, C Xu, S Xu, L Pang, Z Dong, J Xu - arxiv preprint arxiv:2404.11457, 2024 - arxiv.org
With the rapid advancement of large language models (LLMs), information retrieval (IR)
systems, such as search engines and recommender systems, have undergone a significant …

Can large language models be good path planners? a benchmark and investigation on spatial-temporal reasoning

M Aghzal, E Plaku, Z Yao - arxiv preprint arxiv:2310.03249, 2023 - arxiv.org
Large language models (LLMs) have achieved remarkable success across a wide spectrum
of tasks; however, they still face limitations in scenarios that demand long-term planning and …

Cryptotrade: A reflective llm-based agent to guide zero-shot cryptocurrency trading

Y Li, B Luo, Q Wang, N Chen, X Liu… - Proceedings of the 2024 …, 2024 - aclanthology.org
Abstract The utilization of Large Language Models (LLMs) in financial trading has primarily
been concentrated within the stock market, aiding in economic and financial decisions. Yet …

Can llms reason like humans? assessing theory of mind reasoning in llms for open-ended questions

M Amirizaniani, E Martin, M Sivachenko… - Proceedings of the 33rd …, 2024 - dl.acm.org
Theory of mind (ToM) reasoning involves understanding that others have intentions,
emotions, and thoughts, which is crucial for regulating one's reasoning. Although large …

Say more with less: Understanding prompt learning behaviors through gist compression

X Li, Z Liu, C **ong, S Yu, Y Yan, S Wang… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) require lengthy prompts as the input context to produce
output aligned with user intentions, a process that incurs extra costs during inference. In this …

Enhancing LLM trading performance with fact-subjectivity aware reasoning

Q Wang, Y Gao, Z Tang, B Luo, B He - arxiv preprint arxiv:2410.12464, 2024 - arxiv.org
While many studies prove more advanced LLMs perform better on tasks such as math and
coding, we notice that in cryptocurrency trading, stronger LLMs work worse than weaker …

InstructEd: Soft-Instruction Tuning for Model Editing with Hops

X Han, R Li, X Li, J Liang, Z Zhang… - Findings of the …, 2024 - aclanthology.org
The task of model editing becomes popular for correcting inaccurate or outdated parametric
knowledge in Large Language Models (LLMs). However, there are major limitations of state …

CoLE: A collaborative legal expert prompting framework for large language models in law

B Li, S Fan, S Zhu, L Wen - Knowledge-Based Systems, 2025 - Elsevier
Abstract Large Language Models (LLMs) have achieved remarkable outcomes in various
natural language processing tasks. However, their application to the highly specialized field …

GeoHard: Towards Measuring Class-wise Hardness through Modelling Class Semantics

F Cai, X Zhao, H Zhang, I Gurevych… - Findings of the …, 2024 - aclanthology.org
Recent advances in measuring hardness-wise properties of data guide language models in
sample selection within low-resource scenarios. However, class-specific properties are …