Privacy preserving prompt engineering: A survey
Pre-trained language models (PLMs) have demonstrated significant proficiency in solving a
wide range of general natural language processing (NLP) tasks. Researchers have …
wide range of general natural language processing (NLP) tasks. Researchers have …
Information flow control in machine learning through modular model architecture
In today's machine learning (ML) models, any part of the training data can affect the model
output. This lack of control for information flow from training data to model output is a major …
output. This lack of control for information flow from training data to model output is a major …
Instruction Fine-Tuning: Does Prompt Loss Matter?
M Huerta-Enochian, S Ko - … of the 2024 Conference on Empirical …, 2024 - aclanthology.org
We present a novel study analyzing the effects of various prompt loss token weights (PLW)
for supervised instruction fine-tuning (SIFT). While prompt-masking (PLW= 0) is common for …
for supervised instruction fine-tuning (SIFT). While prompt-masking (PLW= 0) is common for …
Confusedpilot: Confused deputy risks in rag-based llms
Retrieval augmented generation (RAG) is a process where a large language model (LLM)
retrieves useful information from a database and then generates the responses. It is …
retrieves useful information from a database and then generates the responses. It is …
MeMemo: On-device Retrieval Augmentation for Private and Personalized Text Generation
Retrieval-augmented text generation (RAG) addresses the common limitations of large
language models (LLMs), such as hallucination, by retrieving information from an updatable …
language models (LLMs), such as hallucination, by retrieving information from an updatable …
DP-RDM: Adapting Diffusion Models to Private Domains Without Fine-Tuning
Text-to-image diffusion models have been shown to suffer from sample-level memorization,
possibly reproducing near-perfect replica of images that they are trained on, which may be …
possibly reproducing near-perfect replica of images that they are trained on, which may be …
Permissive Information-Flow Analysis for Large Language Models
Large Language Models (LLMs) are rapidly becoming commodity components of larger
software systems. This poses natural security and privacy problems: poisoned data retrieved …
software systems. This poses natural security and privacy problems: poisoned data retrieved …
How To Think About End-To-End Encryption and AI: Training, Processing, Disclosure, and Consent
End-to-end encryption (E2EE) has become the gold standard for securing communications,
bringing strong confidentiality and privacy guarantees to billions of users worldwide …
bringing strong confidentiality and privacy guarantees to billions of users worldwide …
DOMBA: Double Model Balancing for Access-Controlled Language Models via Minimum-Bounded Aggregation
The utility of large language models (LLMs) depends heavily on the quality and quantity of
their training data. Many organizations possess large data corpora that could be leveraged …
their training data. Many organizations possess large data corpora that could be leveraged …
Instruction Fine-Tuning: Does Prompt Loss Matter?
M Huerta-Enochian, SY Ko - arxiv preprint arxiv:2401.13586, 2024 - arxiv.org
We present a novel study analyzing the effects of various prompt loss token weights (PLW)
for supervised instruction fine-tuning (SIFT). While prompt-masking (PLW= 0) is common for …
for supervised instruction fine-tuning (SIFT). While prompt-masking (PLW= 0) is common for …