Skip to content

Commit

Permalink
first public commit
Browse files Browse the repository at this point in the history
  • Loading branch information
hypnopump committed Jan 23, 2018
1 parent afdedb9 commit 0ec14ae
Show file tree
Hide file tree
Showing 63 changed files with 11 additions and 147,812 deletions.
51 changes: 11 additions & 40 deletions README.md
Original file line number Diff line number Diff line change
@@ -1,58 +1,29 @@
# E-swish: A New Novel Activation Function
Code for reproducibility of the experiments of [this paper](https://github.com/EricAlcaide/E-swish/blob/master/publishing/ABSTRACT.pdf)
# E-swish: Adjusting Activations to Different Network Depths
This repository will contain the code for reproducibility of the experiments of [this paper](https://arxiv.org/abs/1801.07145v1)

Uploading soon.

## Formula
<div style="text-align:center"><img src ="e_swish.PNG" /></div>

## Abstract

<p align="justify">
Activation functions have a notorious impact in neural networks on both training and testing the models against the desired problem. Currently, the most used activation function is the Rectified Linear Unit (ReLU). Although some alternatives have been proposed, none of them have managed to replace ReLU as the default activation due to inconstant improvements. This paper introduces a new activation function, closely related with the new activation Swish = x * sigmoid (x) (Ramachandran et al., 2017) which we call E-swish.
E-swish is just a Swish activation function mirrored across the identity for all positive values. We show that E-swish outperforms many other well-known activations on a variety of tasks and it also leads to a faster convergence. For example, replacing Relu by E-swish provided at 0.3% accuracy improvement on Cifar10 for the WRN 16-4.

</p>

## Results

### MNIST

| Activations | % Accuracy (median of 3 runs) |
| ------------- | -------------:|
| Leaky Relu (0.3) | 97.71 |
| Elu | 97.92 |
| Relu | 98.30 |
| Swish | 98.31 |
| **E-swish** | **98.46** |

### CIFAR10

* **Simple CNN**

| Activations | % Accuracy (median of 3 runs) |
| ------------- | -------------:|
| Elu | 81.81 |
| Swish | 81.88 |
| Relu | 82.06 |
| Leaky Relu (0.3) | 82.22 |
| **E-swish** | **83.35** |

* **Deeper CNN**

| Activations | % Test error (median of 3 runs) |
| ------------- | -------------:|
| Leaky Relu (0.3) | 8.62 |
| Swish | 8.27 |
| Relu | 8.13 |
| Elu | 7.79 |
| **E-swish** | **7.61** |

<i>*Results are provided on a single model performance.</i>
* **WRN 10-2**

* **WRN 16-4**
|| Activations | % Accuracy (median of 3 runs) |
| ------------- | -------------:|
| Swish | 89.98 |
| Relu | 91.52 |
| **E-swish (beta =1.375)** | **91.89** |

| Activations | % Test error (median of 5 runs) |
| ------------- | -------------:|
| Relu | 5.02 |
| **E-swish** | **4.71** |

## Contact

Expand Down
1,608 changes: 0 additions & 1,608 deletions cifar/Deeper_CNN/.ipynb_checkpoints/cifar10_deeper_cnn_elu-checkpoint.ipynb

This file was deleted.

1,534 changes: 0 additions & 1,534 deletions cifar/Deeper_CNN/.ipynb_checkpoints/cifar10_deeper_cnn_leaky-checkpoint.ipynb

This file was deleted.

Loading

0 comments on commit 0ec14ae

Please sign in to comment.