-
Backpropagation is really an important and fundamental topic in deep learning
yeah, i admit that bp is a little math-heavy and a little hard for new guys
but i also can not imagine that a guy w…
-
https://github.com/Jutho/SparseArrayKit.jl
## N-D tensors
Synapses are by nature 2D (pre-, post-synaptic addresses). In case of multiple spatial dimensions, no reason to burden the synapse matri…
-
First of all, thank you for the comprehensive code base for all variants of S4 models.
However, as I try to run the Listops experiments with S4 (HYYT version), the losses for train, test and val al…
-
1. Das SOM Panel braucht auch die Konsole, die fehlt momentan, lass das panel drin wie vorher.
2. Wenn man von SOM auf Backprop umschaltet, gibt es keine Jung View mehr
3. In de Sidebar muss das Feld …
-
Great work. Thanks for your sharing.
When I use BlurPool, it only needs to change caffe.proto and replace the original base_conv_layer.cpp? And you mentioned that Caffe uses zero paddings instead of …
-
Currently the Neuron.train() method uses the Neuron's error to calculate the delta (used to calculate the gradient and update the weights). This works OK for toy networks with a single output Nueuron…
-
Hi all,
First of all big thanks for the hard work!
I was wondering in the `nerfacto` semantic head if there would be a gradient issue. Indeed in the following line the gradient is detached prev…
-
Hi, Janne
The GatherLayer module in gather.py is smart and efficient. I really appreciate this module.
I almost understand most of the functions of the codes but still confuse at one point.
…
-
Hi Bertinetto,
When you define the shared siamese net, in the backpropagation process.
When it reach the first BatchNorm, after the norm, it will set obj.moments=[] (See matconvnet/matlab/+dagnnn/…
yupup updated
7 years ago
-
The attention formula has an [unusual line](https://github.com/xai-org/grok-1/blob/d6d9447e2d3c9bd81da571dc8681ea60009e4c03/model.py#L865) that puts attention weights through `tanh`.
1. What is the…