A critical review of large language models: Sensitivity, bias, and the path toward specialized ai

A Hajikhani, C Cole - Quantitative Science Studies, 2024 - direct.mit.edu
This paper examines the comparative effectiveness of a specialized compiled language
model and a general-purpose model like OpenAI's GPT-3.5 in detecting SDGs within text …

Contrastive learning with adversarial examples

CH Ho, N Nvasconcelos - Advances in Neural Information …, 2020 - proceedings.neurips.cc
Contrastive learning (CL) is a popular technique for self-supervised learning (SSL) of visual
representations. It uses pairs of augmentations of unlabeled training examples to define a …

Race: Robust adversarial concept erasure for secure text-to-image diffusion model

C Kim, K Min, Y Yang - European Conference on Computer Vision, 2024 - Springer
In the evolving landscape of text-to-image (T2I) diffusion models, the remarkable capability
to generate high-quality images from textual descriptions faces challenges with the potential …

The geometry of feature space in deep learning models: a holistic perspective and comprehensive review

M Lee - Mathematics, 2023 - mdpi.com
As the field of deep learning experiences a meteoric rise, the urgency to decipher the
complex geometric properties of feature spaces, which underlie the effectiveness of diverse …

Improving adversarial robustness through the contrastive-guided diffusion process

Y Ouyang, L **e, G Cheng - International Conference on …, 2023 - proceedings.mlr.press
Synthetic data generation has become an emerging tool to help improve the adversarial
robustness in classification tasks, since robust learning requires a significantly larger …

Ipmix: Label-preserving data augmentation method for training robust classifiers

Z Huang, X Bao, N Zhang, Q Zhang… - Advances in …, 2023 - proceedings.neurips.cc
Data augmentation has been proven effective for training high-accuracy convolutional
neural network classifiers by preventing overfitting. However, building deep neural networks …

Analysis and extensions of adversarial training for video classification

KA Kinfu, R Vidal - … of the IEEE/CVF Conference on …, 2022 - openaccess.thecvf.com
Adversarial training (AT) is a simple yet effective defense against adversarial attacks to
image classification systems, which is based on augmenting the training set with attacks that …

Label noise in adversarial training: A novel perspective to study robust overfitting

C Dong, L Liu, J Shang - Advances in Neural Information …, 2022 - proceedings.neurips.cc
We show that label noise exists in adversarial training. Such label noise is due to the
mismatch between the true label distribution of adversarial examples and the label inherited …

Adversarial unlearning: Reducing confidence along adversarial directions

A Setlur, B Eysenbach, V Smith… - Advances in Neural …, 2022 - proceedings.neurips.cc
Supervised learning methods trained with maximum likelihood objectives often overfit on
training data. Most regularizers that prevent overfitting look to increase confidence on …

[PDF][PDF] Defending against adversarial patches with robust self-attention

N Mu, D Wagner - … on uncertainty and robustness in deep …, 2021 - people.eecs.berkeley.edu
We introduce a new defense against adversarial patch attacks based on our proposed
Robust Self-Attention (RSA) layer. Robust Self-Attention replaces the outlier-sensitive …