-
Hi, I encountered an issue when loading a pre-trained model in the Part-Aware-Transformer repository. Specifically, when trying to resize the position embedding in the Vision Transformer (ViT), I get …
fldzc updated
3 weeks ago
-
This is an excellent work, when will the code be open sourced?
-
Dear RoyiRa:
I try to test whether the LocalBlend can work so I use the run code as following:
`import os
import torch
from prompt_to_prompt_pipeline import Prompt2PromptPipeline
from processors …
-
Thanks for this excellent work! But when I use diffusers to load Controlnet:
`controlnet = SD3ControlNetModel.from_pretrained("stabilityai/stable-diffusion-3.5-controlnets-depth", torch_dtype=torch…
-
**Describe the bug**
When using Flash Attention (--use-flash-attention true) to train Qwen2VL model with mixed data (both image and text data), the code will yield the following error
```
[rank0]: …
-
I dont know why whenever i set use_dora = True it always give me this error when i train:
`RuntimeError Traceback (most recent call last)
Cell In[26], line 1
----> 1 tr…
-
## 🐛 Bug
## Please reproduce using our [template Colab](https://colab.research.google.com/drive/1R-dnKipK9LOVV4_oKbuHoq4VvGKGbDnd?usp=sharing) and post here the link
https://colab.research.g…
-
**Describe the bug**
Running the most recent version of the T5 pretraining script out of the box raises a Value Error, particularly in the following line:
```
[rank0]: File "/home/miniconda3/lib/…
-
### OS
Linux
### GPU Library
CUDA 12.x
### Python version
3.10
### Pytorch version
xxxxxxxxxxx
### Model
turboderp/Mistral-7B-instruct-exl2
### Describe the bug
## Warning: Flash Attention…
-
output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^…