A critical review of large language models: Sensitivity, bias, and the path toward specialized ai
A Hajikhani, C Cole - Quantitative Science Studies, 2024 - direct.mit.edu
This paper examines the comparative effectiveness of a specialized compiled language
model and a general-purpose model like OpenAI's GPT-3.5 in detecting SDGs within text …
model and a general-purpose model like OpenAI's GPT-3.5 in detecting SDGs within text …
Contrastive learning with adversarial examples
Contrastive learning (CL) is a popular technique for self-supervised learning (SSL) of visual
representations. It uses pairs of augmentations of unlabeled training examples to define a …
representations. It uses pairs of augmentations of unlabeled training examples to define a …
Race: Robust adversarial concept erasure for secure text-to-image diffusion model
In the evolving landscape of text-to-image (T2I) diffusion models, the remarkable capability
to generate high-quality images from textual descriptions faces challenges with the potential …
to generate high-quality images from textual descriptions faces challenges with the potential …
The geometry of feature space in deep learning models: a holistic perspective and comprehensive review
M Lee - Mathematics, 2023 - mdpi.com
As the field of deep learning experiences a meteoric rise, the urgency to decipher the
complex geometric properties of feature spaces, which underlie the effectiveness of diverse …
complex geometric properties of feature spaces, which underlie the effectiveness of diverse …
Improving adversarial robustness through the contrastive-guided diffusion process
Synthetic data generation has become an emerging tool to help improve the adversarial
robustness in classification tasks, since robust learning requires a significantly larger …
robustness in classification tasks, since robust learning requires a significantly larger …
Ipmix: Label-preserving data augmentation method for training robust classifiers
Z Huang, X Bao, N Zhang, Q Zhang… - Advances in …, 2023 - proceedings.neurips.cc
Data augmentation has been proven effective for training high-accuracy convolutional
neural network classifiers by preventing overfitting. However, building deep neural networks …
neural network classifiers by preventing overfitting. However, building deep neural networks …
Analysis and extensions of adversarial training for video classification
Adversarial training (AT) is a simple yet effective defense against adversarial attacks to
image classification systems, which is based on augmenting the training set with attacks that …
image classification systems, which is based on augmenting the training set with attacks that …
Label noise in adversarial training: A novel perspective to study robust overfitting
We show that label noise exists in adversarial training. Such label noise is due to the
mismatch between the true label distribution of adversarial examples and the label inherited …
mismatch between the true label distribution of adversarial examples and the label inherited …
Adversarial unlearning: Reducing confidence along adversarial directions
Supervised learning methods trained with maximum likelihood objectives often overfit on
training data. Most regularizers that prevent overfitting look to increase confidence on …
training data. Most regularizers that prevent overfitting look to increase confidence on …
[PDF][PDF] Defending against adversarial patches with robust self-attention
We introduce a new defense against adversarial patch attacks based on our proposed
Robust Self-Attention (RSA) layer. Robust Self-Attention replaces the outlier-sensitive …
Robust Self-Attention (RSA) layer. Robust Self-Attention replaces the outlier-sensitive …