Logo

Aggregated Attack Library Documentation

Foolbox Attacks Library

Deep Fool Attack

Definition: A simple implementation of the fast gradient-based adversarial attack. Allows L2 and L∞ order of the norms.

Parameters:

  • Epsilons
  • Model Lower Bound
  • Model Upper Bound
  • Order of the Norm

Fast Gradient Attack

Definition: Second-order gradient-based attack on the logits. 1 The attack calculates an untargeted adversarial perturbation by performing a approximated second order optimization step on the KL divergence between the unperturbed predictions and the predictions for the adversarial perturbation.

Parameters:

  • Epsilons
  • Model Lower Bound
  • Model Upper Bound
  • Order of the Norm
  • Random Start

Basic Iterative Attack

Definition: The basic iterative attack involves making small, incremental changes to the input data to create effective adversarial examples.

Parameters:

  • Epsilons
  • Model Lower Bound
  • Model Upper Bound
  • Order of the Norm
  • Random Start

Additive Gaussian Noise Attack

Definition: Samples gaussian noise with a fixed L2 size. Can include repeatedly adding to samples, clipping, or both.

Parameters:

  • Epsilons
  • Model Lower Bound
  • Model Upper Bound
  • Order of the Norm
  • Attack Types

Additive Uniform Noise Attack

Definition: Samples uniform noise with a fixed order of the norm size. Can include repeatedly adding to samples, clipping, or both.

Parameters:

  • Epsilons
  • Model Lower Bound
  • Model Upper Bound
  • Order of the Norm
  • Attack Types

Inversion Attack

Definition: Creates “negative images” by inverting the pixel values.

Parameters:

  • Epsilons
  • Model Lower Bound
  • Model Upper Bound

Salt & Pepper Noise Attack

Definition: Increases the amount of salt and pepper noise until the input is misclassified.

Parameters:

  • Epsilons
  • Model Lower Bound
  • Model Upper Bound

Contrast Reduction Attack

Definition: Reduces the contrast of the input using a linear/binary search to find the smallest adversarial. perturbation

Parameters:

  • Epsilons
  • Model Lower Bound
  • Model Upper Bound
  • Search Types

CleverHans Attacks Library:

Fast Gradient Method Attack

Definition: Second-order gradient-based attack on the logits. 1 The attack calculates an untargeted adversarial perturbation by performing a approximated second order optimization step on the KL divergence between the unperturbed predictions and the predictions for the adversarial perturbation.

Parameters:

  • Epsilons
  • Order of the Norm

Projected Gradient Descent Attack

Definition: A technique for generating adversarial examples by iteratively perturbing input data along the gradient direction until reaching a point within a specified distance from the original data that maximizes the loss of a target model.

Parameters:

  • Epsilons
  • Epsilon Step Size
  • Order of the Norm (L2, LInf)
  • Attack Iterations

Basic Iterative Method Attack

Definition: A technique for generating adversarial examples by iteratively applying small perturbations to input data in the direction of the gradient of the loss with respect to the input until reaching a point that maximizes the loss of a target model.

Parameters:

  • Epsilon
  • Epsilon Step Size
  • Attack Iterations
  • Order of the Norm

Momentum Iterative Method Attack

Definition: Momentum iterative method adversarial attack is a variant of the basic iterative method that uses a momentum term to smooth the update direction and accelerate convergence towards the optimal adversarial example.

Parameters:

  • Epsilon
  • Epsilon Step Size
  • Attack Iterations
  • Order of the Norm
  • Decay Factor

Madry Et Al Method Attack

Definition: Madry et al adversarial attack is a powerful variant of the projected gradient descent method that uses a stronger constraint on the size of the perturbation and trains the model using adversarial examples generated with this technique to improve its robustness against future attacks.

Parameters:

  • Epsilon
  • Epsilon Step Size
  • Attack Iterations
  • Order of the Norm

SPSA Attack

Definition: Simultaneous Perturbation Stochastic Approximation (SPSA) adversarial attack is a black-box attack technique that estimates the gradient of the loss function using noisy samples and updates the input data with a step size that is adaptively adjusted to maximize the loss of the target model.

Parameters:

  • Epsilon
  • Model Lower Bound
  • Model Upper Bound
  • Order of the Norm

ML Privacy Meter Attacks Library:

Population Attack

Definition: Attack that determines the ability of attackers to determine if a member is part of a machine learning model training dataset.

Parameters:

  • Loss Function
  • Number of classes