-
Create a new Python file that implements both L2 regularization and dropout, giving users the flexibility to enable or disable these features by adjusting parameters. Ensure all relevant parameters ar…
-
Hello,
I have two questions:
1) The first one is related to the implementation of the 'regularize_kkt()' function. The diagonal is saved and updated with the regularized diagonal at these lines:…
-
Previous works like switch transformers say they use higher dropout rate in the expert modules.
Does OLMoE apply higher expert dropouts (e.g., 0.3 or 0.4) when training?
Taehyun
-
I've been going through your code and don't see an easy way to add it. I looked at the sd_scripts and see there's [this code](https://github.com/kohya-ss/sd-scripts/blob/b8896aad400222c8c4441b217fda0f…
-
How are the levels of regularization for ranking chosen (in the "training" and "test" data sets).
Or in other words, is it important that the submitted method converges quickly for a wide
range of (…
-
I am encountering a problem with the regularization process when using two separate training datasets. Here’s the scenario:
I have two training datasets, each consisting of 10 images.
I also have a …
-
I have some questions about inverse depth. In your paper, I read that both monocular estimation and SFM depth are represented as inverse depth. However, in the code, the inverse depth for SFM is calcu…
-
Should Iterative sense have regularization parameters in the `__call__` ?
Or, alternatively, should this be a separate Reconstruction class, called RegularizedIterativeSENSE?
@fzimmermann89 and @c…
-
I have two questions regarding the dataset used for Appearance Regularization (specified with --reg_data during training).
1. How did you create the dataset? Could you please provide details about th…
-
Hi @JanaldoChen @woo1 et al.,
may i ask if there is a mismatch of background regularization implementations between `exp` in the code
https://github.com/JanaldoChen/Anim-NeRF/blob/1342a9e2d02411…