This document summarizes research on measuring uncertainty in neural networks to detect adversarial examples. It discusses common methods for generating adversarial examples like basic iterative method, fast gradient method, and momentum iterative method. It also explores measures of uncertainty like softmax variance and Bayesian neural networks with Monte Carlo dropout. The researchers tested these uncertainty metrics on MNIST and Kaggle ASSIRA datasets and found dropout-based measures had higher AUC for detecting adversarial examples compared to non-probabilistic neural networks.
Related topics: