-
## lr-sheduler argument reporting errors when fine-tuning mBART
So I'm trying to follow a tutorial to fine tune a pre-trianed model by loading it and then using train.py, below is the command line …
-
Hi, thank you for your good work, I have a question, In paper image2styleGAN++, the author mentioned that they both optimize w and n(noise) , but in your code, I only find w, and find nothing about no…
-
Hi @vietnguyen91 , thanks for the great repo.
I was looking at the plots you have shown for the VOC training and I see a lot of overfitting in the test data. So, I was wondering why is that? Are y…
-
# первый репозиторий на GitHub
import numpy as np
import matplotlib.pyplot as plt
from tensorflow import keras
from tensorflow.keras.layers import Dense
c = np.array([-40, -10, 0, 8, 15, 22, 38])
f …
-
```python
import torch
from torch.autograd import Variable
import torch.optim as optim
l=torch.Tensor([1.01, 1.01, 3, 3])
l=Variable(l.float(), requires_grad = True)
optimizer = optim.Adam([l]…
-
## Description
Running the default command and script for transformer training results in a warning of corrupted index, which is misleading for users (whether the script still works) and should be fi…
-
Hi, I want to train mask rcnn with rle loss!
1. Could you give me some details about the optimizer and learning rate scheduler used in your experiments? I found rle uses the adam while the original m…
-
here is my code:
from tf_unet import unet, util, image_util
#preparing data loading
search_path = 'data/train/*.tif'
data_provider = image_util.ImageDataProvider(search_path)
#setup & train…
-
# principle component with decorrelated
# shuffle the values of principal components:
# each sample now has its
instead of zeroing them out, independently shuffle across all data points, keep the…
-
**Describe the bug**
Deepspeed got segfault when loading CPU_ADAM, both with zero-2 and zero-3 configs / Huggingface transformers integration.
**Zero Configuations**
- Zero-2
```
{
"fp16":…