-
Hi,
is it possible to also upload the training scripts and resulting network weights for the multimodal configuration? (Training on both Optical and Radar data with RandomSensorDrop)
-
After pretraining the model on WebVid, the MSRVTT evaluation results dropped to below 1%. Similarly, when pretraining from the provided pretrained weights, the results also dropped below 1% after the …
-
Hi!
I’ve started developing the Multimodal DataLoader. After taking a (deep) look at this whole multimodal universe, I would like to discuss a couple of things before continuing. I’m using the [to…
-
config_pretrain.yaml
accelerator: gpu
devices: [0, 1]
Hello! Is this only using two graphics cards? How much video memory?
-
While preparing the benchmark for eager and dynamo using the code from the fork: https://github.com/tfogal/NeMo I get errors for dynamo case.
## 🐛 Bug
After fixing [1187](https://github.com/Ligh…
-
Pose a question about one of the following articles:
“[Online images amplify gender bias](https://www.nature.com/articles/s41586-024-07068-x),” 2024. Guilbeault, Douglas, Solène Delecourt, Tasker …
-
Hello! Thanks for the wonderful work and for sharing the pretrained weights.
In your ECCV work (`Making the Most of Text Semantics to Improve Biomedical Vision–Language Processing`), there are seve…
-
Thanks for your great work!
Will the rich feedback model be released?I'd love to test and experience the model and apply it to my own tasks!
-
I admire and am interested in your work and would like to follow up on your work. Will you make the pre-training code and training dataset public?
-
Hello, thank you very much for sharing your work. In the TinyLLaVA_Factory-main file, I executed bash ./scripts/train/train_phi.sh and found a problem. Its loss has been around 5 and has not decreased…