Large language models: a comprehensive survey of its applications, challenges, limitations, and future prospects

MU Hadi, R Qureshi, A Shah, M Irfan, A Zafar… - Authorea …, 2023 - techrxiv.org
Within the vast expanse of computerized language processing, a revolutionary entity known
as Large Language Models (LLMs) has emerged, wielding immense power in its capacity to …

Deep study on autonomous learning techniques for complex pattern recognition in interconnected information systems

Z Amiri, A Heidari, N Jafari, M Hosseinzadeh - Computer Science Review, 2024 - Elsevier
Abstract Artificial Intelligence (AI) and Machine Learning (ML) are being used more and
more to handle complex tasks in many different areas. As a result, interconnected …

Waffling around for performance: Visual classification with random words and broad concepts

K Roth, JM Kim, A Koepke, O Vinyals… - Proceedings of the …, 2023 - openaccess.thecvf.com
The visual classification performance of vision-language models such as CLIP has been
shown to benefit from additional semantic knowledge from large language models (LLMs) …

All in one and one for all: A simple yet effective method towards cross-domain graph pretraining

H Zhao, A Chen, X Sun, H Cheng, J Li - Proceedings of the 30th ACM …, 2024 - dl.acm.org
Large Language Models (LLMs) have revolutionized the fields of computer vision (CV) and
natural language processing (NLP). One of the most notable advancements of LLMs is that a …

Zerog: Investigating cross-dataset zero-shot transferability in graphs

Y Li, P Wang, Z Li, JX Yu, J Li - Proceedings of the 30th ACM SIGKDD …, 2024 - dl.acm.org
With the development of foundation models such as large language models, zero-shot
transfer learning has become increasingly significant. This is highlighted by the generative …

Parameter-efficient fine-tuning for pre-trained vision models: A survey

Y **n, S Luo, H Zhou, J Du, X Liu, Y Fan, Q Li… - arxiv preprint arxiv …, 2024 - arxiv.org
Large-scale pre-trained vision models (PVMs) have shown great potential for adaptability
across various downstream vision tasks. However, with state-of-the-art PVMs growing to …

Text-visual prompting for efficient 2d temporal video grounding

Y Zhang, X Chen, J Jia, S Liu… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
In this paper, we study the problem of temporal video grounding (TVG), which aims to predict
the starting/ending time points of moments described by a text sentence within a long …

Not all prompts are secure: A switchable backdoor attack against pre-trained vision transfomers

S Yang, J Bai, K Gao, Y Yang, Y Li… - Proceedings of the …, 2024 - openaccess.thecvf.com
Given the power of vision transformers a new learning paradigm pre-training and then
prompting makes it more efficient and effective to address downstream visual recognition …

Model reprogramming: Resource-efficient cross-domain machine learning

PY Chen - Proceedings of the AAAI Conference on Artificial …, 2024 - ojs.aaai.org
In data-rich domains such as vision, language, and speech, deep learning prevails to deliver
high-performance task-specific models and can even learn general task-agnostic …

Unifying image processing as visual prompting question answering

Y Liu, X Chen, X Ma, X Wang, J Zhou, Y Qiao… - arxiv preprint arxiv …, 2023 - arxiv.org
Image processing is a fundamental task in computer vision, which aims at enhancing image
quality and extracting essential features for subsequent vision applications. Traditionally …