Skip to content

Feed forward neural network with back propagation

Notifications You must be signed in to change notification settings

dinardo/NeuralNetwork

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

70 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

NeuralNetwork

Feedforward, fully connected, Neural Network with stocastic gradient descent:

  • designed with object oriented paradigm
  • implemented with composition
  • written in python programming language

Multivariate Analysis (MVA.py) implementation with Neural Network
Synopsis: python MVA.py -nv 2 -np 4 -nn 2 20 20 1 -sc

i.e. 2 input variable | 4 perceptrons | 2 - 20 - 20 - 1 neurons | scramble

Check hyper-parameter space before running the program:

  • number of perceptrons and neurons
  • activation function: tanh, sigmoid, ReLU, lin
  • number of mini-batches
  • learn rate, RMSprop, regularization
  • scramble and dropout
  • cost function: quadratic (regression), cross-entropy (classification), softmax

ToDo:

  • add progress bar
  • normalize input variabile: mean = 0, RMS = 1
  • output layer with sigmoid, to go from 0 to 1, and hidden layers with tanh (?)
  • bias in weights
  • softmax for linear classifier cs231n.github.io
  • plot NN output for signal and background
  • plot ROC integral
  • plot F-score

To Check:

  • weight initialization: Gaussian, Uniform
  • implementaion of stocastic gradient descent
  • implementaion of RMSprop
  • if output activation function can be made different: linear (regression), sigmoid (classification), softmax (multi-class classification)

About

Feed forward neural network with back propagation

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages