-
Deep-SAD-PyTorch/src/DeepSAD.py, **line 123**
The encoder has different layer names than the network. As a result, the ae_net_dict here is an empty dictionary. One possible revision,
…
-
I am running the pretraining code the way you suggested but it has been stuck at this point for 2 hours now. Is this supposed to take this long?
```console
neilpaul77@NeilRig77:~/Downloads/ntua-slp-…
-
## Paper Link
https://arxiv.org/abs/2002.01685
https://github.com/aghie/parsing-as-pretraining
## Upload
2020/2/5
## What is paper about?
## Paper Contributions
## Key Points
## Va…
-
Hi,
I am trying to reproduce pretraining of mt5 model, when you modify the sentences as:
`Thank you to week => for inviting me your party last `
Then do you compute the loss on all to…
-
## ❓ Questions and Help
I am trying to pretrain wav2vec2 on persian language using common voice dataset. I did not modify anything but the dataset path in configs. here is plots of training metrics…
-
## ❓ Questions and Help
According to the paper, it only does sentence permutation and span masking (ignoring the other noises that BART is pretrained with). I found [something](https://github.com…
-
### Describe the bug
This issue serves as a placeholder for the introduction of two recipes: full whisper pretraining and a speechLLM finetuning.
People potentially interested are added.
Anyone …
-
![image](https://github.com/user-attachments/assets/6b7fa225-4120-450a-a8a9-aa259100b40f)
However, the whole 50M cells can't be downloaded with only these scripts.
Thanks
-
When trying to pretrain t5-base, we are seeing that that pretraining loss starts at an enormous number (~160000).
Even when trying to pretrain smaller variants of t5, the initial pretraining loss alw…
-
Really nice work! I have questions about the pretraining of ViT. The paper mentions that the ViT is pretrained on publicly available histology datasets through SSL. I am wondering where I can find the…