-
Hi author, I am having some problems trying to reproduce the DaiSEE accuracy of 63.9% from your paper, could you please help me, it won't take much of your time.
1. i have added weighted sampling in …
-
您好!我想请问一下,在您的论文中我看到选择的是SGD作为Optimizer,我在使用的时候尝试使用Adam和AdamW Optimizer,但是没训练几轮loss就变成了NAN,不知道您这边有没有遇到过类似的问题,还是说就只是使用了SGD呢?
-
There are 4 same optimizer,why?
raw_parameters = list(net.pretrained_model.parameters())
part_parameters = list(net.proposal_net.parameters())
concat_parameters = list(net.concat_net.parameters()…
-
I am encountering issues when using non-element-wise optimizers such as Adam-mini with DeepSpeed.
According to the documentation, it reads:
> The FP16 Optimizer is designed to maximize the achievable…
-
It seems like the first call to `step` of the `GradientDescent` optimizer doesn't perform the step operation. I didn't check if this occurs for other optimizers or do other digging, but can do so if t…
-
SGD, Cross Entropy als Beispiel implementieren (ggf. nur weiterer Import nötig)
-
When I'm trying to run the esn_sgd code, the optimizer is giving Optimizer is getting empty list error
-
## ❓ Questions and Help
Hi All,
I Have this code
```
import optuna
from torch.optim.lr_scheduler import ReduceLROnPlateau
# Assuming dataset is already defined
train_size = int(0.8 * len(da…
-
### 🚀 Feature
Hello all,
I would like to add Adabound to the list of existing optimizers in the torch.optim module.
Here is the link to the paper - https://openreview.net/pdf?id=Bkg3g2R9FX
…
-
When I try to trian resnet50 by dino on on X-Ray dataset with 1000 images, the loss is not drop, even increase some time. Have anyone met the same issue and solve it.
the config as follow:(at first, …