dicarlolab / archconvnets

Architecturally optimized neural networks trained with regularized backpropagation
5 stars 5 forks source link

Modified code to allow caching, added example module for loading #21

Open ardila opened 10 years ago

ardila commented 10 years ago

@yamins81 This solution seems to work. I tried something more general at first, but this seems to work without requiring too much modificatino of the interface. I had to change a couple of things:

  1. passing the model as an argument to the weight and bias initialization functions so that I can cache stored models and not have to reload them for every layer
  2. allow for weight and bias intialization functions to return a tuple: first value is used to intialize, second is used to initialize the increment.