-
Thinking of reading this:
https://www.amazon.com/Irresistible-Addictive-Technology-Business-Keeping/dp/1594206643
(After Cal Newport's Digital Minimalism, of course!)
-
Hi. This is not a issue, but a question. I suspect you compared several optimizers for SGD before ended with adam. Was it better than others? I also have FM implementation which uses adagrad so far. …
-
**Objective**:
Evaluate the stability of the grassland notebook featuring a GF map and assess its suitability for sharing with other projects. If deemed unstable, adapt it to provide a more stable exa…
-
### 🐛 Describe the bug
When running with Adam eager, about 1/3 of our benchmark models fail accuracy. It's about uniform across suites.
[list of failing models](https://github.com/pytorch/pytorch/…
-
Thank you for taking the time to complete our test and to agree to provide some improvements.
This is entire part of the interview process.
There are no time constraints from my side, so you can p…
-
### 🐛 Describe the bug
Adam optimizer seems to be leaking memory(?) in AMD chips.
A minimal example doing some simple training can be found here:
https://gist.github.com/lingchunkai/dbb0c001a2fc…
-
-
Seems AdamWeightDecay is not a general Adam implement with weightdecay, it's much different from pytorch's [Adam](https://pytorch.org/docs/stable/_modules/torch/optim/adam.html#Adam).
Pytorch's decay…
-
-
**Describe the bug**
I installed deepspeed with pip install deepspeed and tried to use DeepSpeedCPUAdam but with this error
```
Exception ignored in:
Traceback (most recent call last):
File …