-
Hi everyone!
Thanks for this great work. Since several months, MAE is available in HuggingFace Transformers with an easy-to-use and familiar API: https://huggingface.co/docs/transformers/model_doc/…
-
Thank you @v-nhandt21 for sharing the repo. I have two questions, if you have time, please help me.
1. Before `script` is transformed into `phoneme` through the function `vi2IPA_split`, does it nee…
-
I am trying to train a model on my macbook (M1 Max) and even after 30k steps, it still sounds very noisy. The loss seems to converge (very slowly), but it still sounds unbearable.
Original Post:
…
-
Hi,
I have Image and Description of Products which is in Spanish language and want to train a classifier model using ViLT. What kind of pretrained model shall I use to train it on my Spanish Text and…
-
### Question
Hi,
I have successfully pretrained the mm_projector, and finish the finetune stage with following script:
```
################## LLaMA-2 ##################
PROMPT_VERSION="lla…
-
Hello! I'm trying to reproduce the result on CC-CCII and MM-WHS.
- For CC-CCII, what is the training-validation-testing setting? Is it a 3-fold cross-validation?
- For MM-WHS, is the testing don…
-
### Describe the bug
no idea what's the bug but when I type svc pre-resample, it doesn't work.
I type these commands again to see if there isn't a download that isn't working properly but it says re…
-
[HiFiGan](https://github.com/jik876/hifi-gan) has sota results in wav generation from mel spectrograms
Is it possibile to add support to `hifigan` model, after the `mel` generation, in order to…
-
Hi,
Wondering if you could upload trained models for different embedding sizes?
Thanks
-
Hi,
First of all, thanks for the great work!
### Issue I encountered:
I am trying to replicate the BLIP-2 paper, Table3,
![Screenshot 2024-05-29 at 23 36 23](https://github.com/salesforce/LAV…