Models trained with various filters; kuwahara filter defense
This commit is contained in:
36
Filter_Analysis/wiki/DesignImpact.md
Normal file
36
Filter_Analysis/wiki/DesignImpact.md
Normal file
@ -0,0 +1,36 @@
|
||||
# Engineering Design Principles
|
||||
1. Clearly defined problem
|
||||
- Assess the efficacy of various denoising filters in preserving the accuracy of image classifier models under a noise-based attack.
|
||||
2. Requirements
|
||||
- Only algorithmic approach for defense
|
||||
- Must be faster than auto-encoder
|
||||
3. Constraints
|
||||
- Computing power
|
||||
- Memory usage
|
||||
- Impossible to know who and how a model will be attacked
|
||||
4. Engineering standards
|
||||
- [[https://peps.python.org/pep-0008/|PEP 8]]
|
||||
-
|
||||
5. Cite applicable references
|
||||
- [[https://pytorch.org/tutorials/beginner/fgsm_tutorial.html|FGSM Attack]]
|
||||
- [[https://github.com/pytorch/examples/blob/main/mnist/main.py|MNIST Model]]
|
||||
- [[https://www.cs.toronto.edu/~kriz/cifar.html|CIFAR-10]]
|
||||
6. Considered alternatives
|
||||
a) Iterate on the design
|
||||
i) Advantages
|
||||
- Potentially more computationally efficient than an ML approach
|
||||
- Will likely use less memory than a model used to clean inputs
|
||||
- No training (very computationally intense) stage
|
||||
ii) Disadvantages
|
||||
- Potentially less effective than than an ML approach
|
||||
iii) Risks
|
||||
- Conventional algorithm may be more vulnerable to reverse engineering
|
||||
7. Evaluation process
|
||||
- Cross validation
|
||||
- Effectiveness will be measured as the percent of correct classifications
|
||||
- Testing clean vs. filtered training data
|
||||
- Ablation variables:
|
||||
- Different models
|
||||
- Different datasets
|
||||
- Different filters
|
||||
8. Deliverables and timeline
|
@ -4,5 +4,5 @@
|
||||
- [[Tests]]
|
||||
- [[Approach]]
|
||||
- [[Rationale]]
|
||||
- [[Notes]]
|
||||
- [[DesignImpact]]
|
||||
- [[Timeline]]
|
||||
|
42
Filter_Analysis/wiki/Results.wiki
Normal file
42
Filter_Analysis/wiki/Results.wiki
Normal file
@ -0,0 +1,42 @@
|
||||
= Experimental Results =
|
||||
|
||||
== Model Trained on Unfiltered MNIST Dataset ==
|
||||
| Epsilon | Accuracy |
|
||||
|---------|----------|
|
||||
| 0.05 | 0.9600 |
|
||||
| 0.10 | 0.8753 |
|
||||
| 0.15 | 0.7228 |
|
||||
| 0.20 | 0.5008 |
|
||||
| 0.25 | 0.2922 |
|
||||
| 0.30 | 0.1599 |
|
||||
|
||||
== Model Trained on Kuwahara (R=5) Filtered MNIST Dataset ==
|
||||
| Epsilon | Attacked Accuracy | Filtered Accuracy | Ratio |
|
||||
|---------|-------------------|-------------------|--------|
|
||||
| 0.05 | 0.9605 | 0.9522 | 0.9914 |
|
||||
| 0.1 | 0.8743 | 0.9031 | 1.0329 |
|
||||
| 0.15 | 0.7107 | 0.8138 | 1.1451 |
|
||||
| 0.2 | 0.4876 | 0.6921 | 1.4194 |
|
||||
| 0.25 | 0.2714 | 0.5350 | 1.9713 |
|
||||
| 0.3 | 0.1418 | 0.3605 | 2.5423 |
|
||||
|
||||
== Model Trained on Gaussian Blurred (K-Size=5x5) MNIST Dataset ==
|
||||
| Epsilon | Attacked Accuracy | Filtered Accuracy | Ratio |
|
||||
|---------|-------------------|-------------------|-------|
|
||||
| 0.05 | 0.9192 | 0.9325 | 1.014 |
|
||||
| 0.10 | 0.7629 | 0.8802 | 1.154 |
|
||||
| 0.15 | 0.4871 | 0.7865 | 1.615 |
|
||||
| 0.20 | 0.2435 | 0.6556 | 2.692 |
|
||||
| 0.25 | 0.1093 | 0.5024 | 4.596 |
|
||||
| 0.30 | 0.0544 | 0.3522 | 6.474 |
|
||||
|
||||
== Model Trained on Bilateral Filtered (d=5) MNIST Dataset ==
|
||||
| Epsilon | Attacked Accuracy | Filtered Accuracy | Ratio |
|
||||
|---------|-------------------|-------------------|-------|
|
||||
| 0.05 | 0.9078 | 0.9287 | 1.023 |
|
||||
| 0.10 | 0.7303 | 0.8611 | 1.179 |
|
||||
| 0.15 | 0.4221 | 0.7501 | 1.777 |
|
||||
| 0.20 | 0.1927 | 0.6007 | 3.117 |
|
||||
| 0.25 | 0.0873 | 0.4433 | 5.078 |
|
||||
| 0.30 | 0.0525 | 0.3023 | 5.758 |
|
||||
|
Reference in New Issue
Block a user