Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

SNN (Self-Normalizing Networks) #12

Open
1 of 3 tasks
mdraw opened this issue Jul 15, 2017 · 0 comments
Open
1 of 3 tasks

SNN (Self-Normalizing Networks) #12

mdraw opened this issue Jul 15, 2017 · 0 comments

Comments

@mdraw
Copy link
Member

mdraw commented Jul 15, 2017

https://arxiv.org/abs/1706.02515 proposes a new activation function SELU, which can replace explicit batch normalization and could improve overall performance and accuracy of our neural networks.
To do:

  • Implement SELU activation function
  • Implement special "alpha dropout" for SNNs because regular dropout is broken with SELU (see p. 6 of paper). I think we can do this transparently by automatically switching to the alpha dropout algorithm in Perceptron constructors if they detect activation_func='selu'
  • Weight initialisation method suggested at the bottom of p.3 of the paper (→ elektronn2.neuromancer.variables.initweights())

Links:

mdraw added a commit that referenced this issue Jul 15, 2017
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant