Interpretable deep learning: Interpretation, interpretability, trustworthiness, and beyond
Deep neural networks have been well-known for their superb handling of various machine
learning and artificial intelligence tasks. However, due to their over-parameterized black-box …
learning and artificial intelligence tasks. However, due to their over-parameterized black-box …
Sok: Model inversion attack landscape: Taxonomy, challenges, and future roadmap
SV Dibbo - 2023 IEEE 36th Computer Security Foundations …, 2023 - ieeexplore.ieee.org
A crucial module of the widely applied machine learning (ML) model is the model training
phase, which involves large-scale training data, often including sensitive private data. ML …
phase, which involves large-scale training data, often including sensitive private data. ML …
Extracting training data from diffusion models
Image diffusion models such as DALL-E 2, Imagen, and Stable Diffusion have attracted
significant attention due to their ability to generate high-quality synthetic images. In this work …
significant attention due to their ability to generate high-quality synthetic images. In this work …
Fine-tuning aligned language models compromises safety, even when users do not intend to!
Optimizing large language models (LLMs) for downstream use cases often involves the
customization of pre-trained LLMs through further fine-tuning. Meta's open release of Llama …
customization of pre-trained LLMs through further fine-tuning. Meta's open release of Llama …
Diffusion art or digital forgery? investigating data replication in diffusion models
Cutting-edge diffusion models produce images with high quality and customizability,
enabling them to be used for commercial art and graphic design purposes. But do diffusion …
enabling them to be used for commercial art and graphic design purposes. But do diffusion …
Text embeddings by weakly-supervised contrastive pre-training
This paper presents E5, a family of state-of-the-art text embeddings that transfer well to a
wide range of tasks. The model is trained in a contrastive manner with weak supervision …
wide range of tasks. The model is trained in a contrastive manner with weak supervision …
Beyond neural scaling laws: beating power law scaling via data pruning
Widely observed neural scaling laws, in which error falls off as a power of the training set
size, model size, or both, have driven substantial performance improvements in deep …
size, model size, or both, have driven substantial performance improvements in deep …
D4: Improving llm pretraining via document de-duplication and diversification
Over recent years, an increasing amount of compute and data has been poured into training
large language models (LLMs), usually by doing one-pass learning on as many tokens as …
large language models (LLMs), usually by doing one-pass learning on as many tokens as …
Quantifying memorization across neural language models
Large language models (LMs) have been shown to memorize parts of their training data,
and when prompted appropriately, they will emit the memorized training data verbatim. This …
and when prompted appropriately, they will emit the memorized training data verbatim. This …
Data-efficient Fine-tuning for LLM-based Recommendation
Leveraging Large Language Models (LLMs) for recommendation has recently garnered
considerable attention, where fine-tuning plays a key role in LLMs' adaptation. However, the …
considerable attention, where fine-tuning plays a key role in LLMs' adaptation. However, the …