Privacy preserving prompt engineering: A survey

K Edemacu, X Wu - arxiv preprint arxiv:2404.06001, 2024 - arxiv.org
Pre-trained language models (PLMs) have demonstrated significant proficiency in solving a
wide range of general natural language processing (NLP) tasks. Researchers have …

Information flow control in machine learning through modular model architecture

T Tiwari, S Gururangan, C Guo, W Hua… - 33rd USENIX Security …, 2024 - usenix.org
In today's machine learning (ML) models, any part of the training data can affect the model
output. This lack of control for information flow from training data to model output is a major …

Instruction Fine-Tuning: Does Prompt Loss Matter?

M Huerta-Enochian, S Ko - … of the 2024 Conference on Empirical …, 2024 - aclanthology.org
We present a novel study analyzing the effects of various prompt loss token weights (PLW)
for supervised instruction fine-tuning (SIFT). While prompt-masking (PLW= 0) is common for …

Confusedpilot: Confused deputy risks in rag-based llms

A RoyChowdhury, M Luo, P Sahu, S Banerjee… - arxiv preprint arxiv …, 2024 - arxiv.org
Retrieval augmented generation (RAG) is a process where a large language model (LLM)
retrieves useful information from a database and then generates the responses. It is …

MeMemo: On-device Retrieval Augmentation for Private and Personalized Text Generation

ZJ Wang, DH Chau - Proceedings of the 47th International ACM SIGIR …, 2024 - dl.acm.org
Retrieval-augmented text generation (RAG) addresses the common limitations of large
language models (LLMs), such as hallucination, by retrieving information from an updatable …

DP-RDM: Adapting Diffusion Models to Private Domains Without Fine-Tuning

J Lebensold, M Sanjabi, P Astolfi… - arxiv preprint arxiv …, 2024 - arxiv.org
Text-to-image diffusion models have been shown to suffer from sample-level memorization,
possibly reproducing near-perfect replica of images that they are trained on, which may be …

Permissive Information-Flow Analysis for Large Language Models

SA Siddiqui, R Gaonkar, B Köpf, D Krueger… - arxiv preprint arxiv …, 2024 - arxiv.org
Large Language Models (LLMs) are rapidly becoming commodity components of larger
software systems. This poses natural security and privacy problems: poisoned data retrieved …

How To Think About End-To-End Encryption and AI: Training, Processing, Disclosure, and Consent

M Knodel, A Fábrega, D Ferrari, J Leiken… - arxiv preprint arxiv …, 2024 - arxiv.org
End-to-end encryption (E2EE) has become the gold standard for securing communications,
bringing strong confidentiality and privacy guarantees to billions of users worldwide …

DOMBA: Double Model Balancing for Access-Controlled Language Models via Minimum-Bounded Aggregation

T Segal, A Shabtai, Y Elovici - arxiv preprint arxiv:2408.11121, 2024 - arxiv.org
The utility of large language models (LLMs) depends heavily on the quality and quantity of
their training data. Many organizations possess large data corpora that could be leveraged …

Instruction Fine-Tuning: Does Prompt Loss Matter?

M Huerta-Enochian, SY Ko - arxiv preprint arxiv:2401.13586, 2024 - arxiv.org
We present a novel study analyzing the effects of various prompt loss token weights (PLW)
for supervised instruction fine-tuning (SIFT). While prompt-masking (PLW= 0) is common for …