-
# Paper Review: CogVLM: Visual Expert for Pretrained Language Models – Andrey Lukyanenko
My review of the paper CogVLM Visual Expert for Pretrained Language Models
[https://andlukyane.com/blog/paper…
-
```
[2024-03-20 16:15:45,873] [INFO] [real_accelerator.py:110:get_accelerator] Setting ds_accelerator to cuda (auto detect)
config.json: 100%|████████████████████████████████████████████████████████…
-
Hi,
Thank you for this wonderful work!
When I was trying to reproduce your results, I faced some challenges when getting a minimum working example to run.
```python
import __main__
imp…
-
(clip4str) root@Lab-PC:/workspace/Project/OCR/CLIP4STR# bash scripts/vl4str_base.sh
abs_root: /home/shuai
model:
_convert_: all
img_size:
- 224
- 224
max_label_length: 25
charset_t…
-
Hi,
I hope you are doing fine actually i m confused in one thing regarding the CLIP-G models because as your training data looks more of like G prompt and not L prompt which I like comma separated th…
-
### Checklist
- [X] The issue has not been resolved by following the [troubleshooting guide](https://github.com/lllyasviel/Fooocus/blob/main/troubleshoot.md)
- [ ] The issue exists on a clean install…
-
I have noticed that the contextbart model is defined in the repository but I did not see any script in the scripts folder that explicitly calls or uses this model. Could you please provide guidance on…
-
We have trained StyleTTS2 model for Hindi language. Initially we trained PL-bert for Hindi considering Espeak phonemizer and Indicbert tokenizer. Then we utilized that newly trained Hindi PLbert by re…
-
Hi
I discovered this project and is pretty amazing the results provided, I saw that it got updated to get support for the Japanese language, and that gave me the curiosity of how many epochs or hou…
-
Is the fine-tuning of InternVL supported by hugging face SFTTrainer?
I got the following error when using the SFTTrainer:
```python
model = AutoModel.from_pretrained(
"OpenGVLab/…