Skip to content

Fast Gradient Sign Adversarial Attack(FGSM) examples creation using FashionMnist dataset

Notifications You must be signed in to change notification settings

jaydeepthik/Fast-Gradient-Sign-Adversarial-Attack

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 

Repository files navigation

Fast-Gradient-Sign-Adversarial-Attack

Fast Gradient Sign Adversarial Attack(FGSM) examples creation using FashionMnist dataset.
inspiration: https://pytorch.org/tutorials/beginner/fgsm_tutorial.html

About Adversarial Attacks:

Fast Gradient Sign Attack:

General goal: Add the least amount of pertubations to the input that causes desired misclassification.

Assumptions on attacker's knowledge:

  1. White-Box: Attacker has full knowledge and access to the model, architecture, inputs, outputs and weights.
  2. Black-Box: Attacher has knowledge only about the inputs and outputs of the model and no information about the underlynig model architecture or weights.

Goals:

  1. misclassification: Attacker only wants the output classification to be wrong and does not care about what the new classification is.
  2. Source/Target misclassification: pertubations to the input that belongs to a specific source class so that it is classified as a specific target class.

FGSM : White-box attack with the goal of misclassification.

Fast Gradient Sign Attack: Use gradient of the loss w.r.t input data, then adjust the inputs to maximize the loss


In this repo we implement FGSM on the FashionMNIST dataset

About

Fast Gradient Sign Adversarial Attack(FGSM) examples creation using FashionMnist dataset

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published