-
你好,我根据readme提供的路径下载了APE-D模型与配置文件
运行脚本如下:
`python demo/demo_lazy.py \
--config-file configs/LVISCOCOCOCOSTUFF_O365_OID_VGR_SA1B_REFCOCO_GQA_PhraseCut_Flickr30k/ape_deta/ape_deta_vitl_eva02_clip_vlf_…
-
Hi! I am exploring sentence transformers for a visual scene detection application, to correct automated close captioning according to what is found in the analyzed video frame. For example, if the vid…
-
I am trying to use this project with a vision-language model like https://huggingface.co/docs/transformers/en/model_doc/llava_next but currently this repo does not support vision part of the model. I …
-
Hi! !مرحبا! السلام عليكم
Let's bring the documentation to all the Arabic-speaking community 🌏 (currently 0 out of 267 complete)
Would you want to translate? Please follow the 🤗 [TRANSLATING guid…
-
This issue tracks the progress on improving the handling and testing of Vision-Language Models. The main goals are to enhance/enable generation tests, handle other generation techniques like assisted …
-
First of all, thank you for your work. I have a question for you.
I want to fine-tune a complete text encoder model, but it seems that the model trained by ft-B-train-OpenAI-CLIP-ViT-L-14.py is a vis…
-
I want to make visualization systems for visualizing transformers, specifically self-attention. It would be nice if it worked for Vision Transformers as well as Language Models.
-
CVPR 2022
#
格式
* **Paper Title**
*Author(s)*
CVPR, 2022. [[Paper]](link) [[Code]](link) [[Website]](link)
需要填充:
1)Paper Title
2) Author(s)
3) 3个“link”
4)两篇文章之间间隔一行
# agent
Meta Ag…
yyf17 updated
2 years ago
-
- [ ] [MoAI/README.md at master · ByungKwanLee/MoAI](https://github.com/ByungKwanLee/MoAI/blob/master/README.md?plain=1)
# MoAI/README.md at master · ByungKwanLee/MoAI
## Description
![MoAI: Mixture…
-
Hello again!
Would it be possible to modify the GMP fine tune script to train a LoRA with PEFT for the CLIP VIT-G model? Then merge the LoRA with the model to get a new CLIP-G model?
Chat-GPT se…