Skip to content

Code for our project on initialization strategies for DNN with maxout activation functions.

Notifications You must be signed in to change notification settings

ZivaUrbancic/Maxout_Initializations

Repository files navigation

Initialization of Deep Neural Networks with ReLU Activation

Linear regions of networks with piecewise linear activation play an important role in their understanding. Since networks with higher number and more even spread of linear regions are believed to be able to approximate a richer class of functions, it may be beneficial to maximize their number. We implement our own initialization strategy with that aim, and run experiments comparing it with more standard strategies. We trained the networks (with 2 hidden layers of 10 units each and ReLU activation) for classification on MNIST data set. The code for experiments with isotropic scaling can be found on isoscale branch.

About

Code for our project on initialization strategies for DNN with maxout activation functions.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •