-
hi ,there
i'm training an AM on 4 TeslaV100 gpus initially using sgd each epoch is done in ~ 12 min, since I'm having some problems with the lr, I wanted to move to novograd optimizer unfortunatly is…
-
File "D:\A_File\Project\RepLKNet-pytorch-main\optim_factory.py", line 19, in
from timm.optim.novograd import NovoGrad
ModuleNotFoundError: No module named 'timm.optim.novograd'
-
Great work!
Can you, please, add implementation for Novograd algorithm?
Support info:
paper: https://arxiv.org/abs/1905.11286
Novograd implementations:
https://github.com/NVIDIA/apex/blob/mas…
-
In both of the jupyter notebooks and the paper, I noticed that instead of using Adam, the most commonly used optimizer for transformers, you used Adagrad for all of the experiments. Is there a reason …
-
any plan on migrating code to tf 2.0?
-
- [x] Auto mixed precision and loss scaling
- [x] Fix transducer embedding tflite conversion (tflite conversion raises a bug when using `tf.gather` in `tf.while_loop`)
- [x] Fix transducer tflite co…
-
self.lr = K.variable(lr, name='lr')
AttributeError: can't set attribute
keras version - 2.3.1
How to solve this issue?
-
Hi ,
Thanks so much for the wrapper API.
I was trying to make real-time KWS spotting using your wrapper and my trained .tflite model using the paper [google paper](https://arxiv.org/pdf/2005.067…
-
Currently, it is not possible to cache data to GPU during training of deepedit, to accelerate training as described in the [Fast Training Tutorial](https://github.com/Project-MONAI/tutorials/blob/main…
-
There are a number of differences compared to the source:
* original article: https://arxiv.org/pdf/1910.10261.pdf
Differences:
1. The default training script uses LibriTTS dataset instead of…