-
Hi,
Thanks for providing the pre-training database with foldseek tokens! having difficulty downloading the dataset and using with hugginface functions. Trying
```
from datasets import load_da…
-
![image](https://github.com/user-attachments/assets/dfce3f7c-ca24-4216-a88f-a96b47cc0be8)
-
Will you guys be adding support for PreTraining ?
-
After pretraining the model on WebVid, the MSRVTT evaluation results dropped to below 1%. Similarly, when pretraining from the provided pretrained weights, the results also dropped below 1% after the …
-
0%| | 0/600000 [00:00
-
Hello! I'm running EEG pretraining with code stageA1_eeg_pretrain.py. I'm wondering do you provide the Pretraining EEG Dataset? I saw the path in this code (../dreamdiffusion/datasets/mne_data/) doesn…
-
I can see the new notebook to train/add a new language, but I don't know why no extra tokens are being added. The new language will not necessarily have the same script or origin. How is it a good way…
-
Hello
I read the code&paper and find that PLM is trainable when in the pretraining stage.Do you try to freeze PLM and only train text projection layer in the pretraining stage? If you have tried, h…
-
We have trained SSL model based on the NEST, how to finetune the previous model based on the CTC Loss function?
The pretraining scripts are as follows: [NEST](https://github.com/NVIDIA/NeMo/blob/main…
-
Hi,
Congrats on your accepted work! I'd have some questions to understand the model architecture and performance better.
- What patch encoder did you use in the CLAM baseline? Is it based on Re…