-
I tried to reproduce the results for VinVL+VIVO+SCST on NoCaps, but my result was off by a visible margin.
### Reported Results on NoCaps validation set
`"CIDEr": {"in-domain": 103.7, "near-domain…
-
Hi Yuan,
Thanks again for this great work, I have been using both this and the original AST model for some downstream tasks. I am currently looking into some other time series data, and was wonder…
-
Hi authors!
Thank you for making the paper and code open source. It is very helpful.
I am trying to pretrain the GDT model on kinetics400 dataset, while I spent more than 1 day on each epoch. I run …
-
Can i use training_multi-task for pretraining model using mlm task and contrastive loss at the same time? My data are all sentences pair.
Looking forward to your reply!
-
Hi Stefan,
When I use the Turkish model on an English dataset for classification, it works surprisingly well. So, I have two questions:
1) Does the training corpus contain English texts?
2) I…
-
Could someone explain to me what exactly this class does?
Is it possible to get the classification output without pretraining?
(It takes too long on colab GPU. I need something I can run on that)
-
Mistral-7b is a much better model (and perhaps a teacher) than Llama-2-7b. Would you kindly release checkpoints for a distilled mistral? Would greatly appreciate it!
ojus1 updated
2 months ago
-
Hi @henryzhongsc, thanks for your work on this repo.
I was wondering whether it could be possible to train a different model than a ResNet20 with your current state of the code.
I am trying to p…
-
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
有看到llama支持continual pretraining,想问一下chatglm能支持这种继续预训练方式吗?
### Expected Behavior
_No respons…
zoepo updated
11 months ago
-