Skip to content

Developed robust image classification models to prevent the effect of adversarial attacks

Notifications You must be signed in to change notification settings

arushi2509/Defense-Mechanisms-Against-Adversarial-Attacks-in-Computer-Vision-

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

44 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Defense Mechanisms Against Adversarial Attacks in Computer Vision

This project focuses on the implementation of a deep convolutional neural network (CNN) using the ResNet18 architecture to classify images from the CIFAR-10 dataset. Additionally, we explored adversarial attacks on the trained model and implemented defense mechanisms to counteract these attacks.

Table of Contents

Dataset

  • CIFAR-10: 60,000 RGB images (32x32), with 10 classes such as frogs, horses, ships, trucks etc.
    • 50,000 training images
    • 10,000 testing images

cifar10

Model

  • Architecture: ResNet18
  • Accuracy: 96% on the training dataset and 80% on the test dataset.

Adversarial Attacks & Defense

Various attacks were implemented on the trained model, and their impact on model performance was observed:

  • Noise Attacks (with varying standard deviations):

    • Stddev. 0.01 - Accuracy Drop: 0%
    • Stddev. 0.09 - Accuracy Drop: 8%
    • Stddev. 0.25 - Accuracy Drop: 32%
    • Stddev. 0.50 - Accuracy Drop: 49%
  • Other Attacks:

    • FGSM (Fast Gradient Sign Attack)
    • PGD (Projected Gradient Descent)
    • C-W (Carlini-Wagner)

Results with different epsilon values:

Epsilon FGSM PGD C-W
0.002 54.36% 54.75% 58.5%
0.02 27.55% 24.31% 57.5%
0.2 6.67% 0.6% 53.5%

Defense

A defense mechanism was implemented against the Additive Gaussian Noise Attack:

  • Defense Strategy: Adversarial training with gaussian noise.
    • Accuracy on adversarial test set increased from 72% to 77%.

Conclusion

The project showcases the performance of ResNet18 on the CIFAR-10 dataset and delves deep into understanding the effects of adversarial attacks. We explored the trade-off between the visual quality of the adversarial image and its impact on model performance. Furthermore, adversarial training proved to be an effective defense mechanism.