-
To save GPU memory, I want to load the multilingual model in 4bit mode, the code is as follows.
```python
import torch
from transformers import AutoTokenizer
from mplug_owl.modeling_mplug_owl impo…
-
Hi! First I just want to say, thank you for making this opensource and doing such an amazing job with this.
I'm having no problems with text2video, but with video2video I'm having problems. Any of …
-
Loading checkpoint shards: 50%|█████████████████████████████████████████████████████████████████████████████████████████ …
-
### Model description
X-Decoder is a generalized decoding pipeline that can predict pixel-level segmentation and language tokens seamlessly. X-Decoder is the first work that provides a unified way to…
-
I want to transform the table detection model from detr to onnx. Some models available in HF are either "nielsr/detr-table-detection" or "microsoft/table-transformer-detection".
I try both and with…
-
**PLEASE DO CHIME IN WITH YOUR THOUGHTS**
My personal idea for this module is:
To have a raku module that does this set of things around HTMX (see [htmx.org](htmx.org)):
- write web pages in …
-
llava multimodel would be huge to be supported for aws neuron chips
https://huggingface.co/llava-hf/llava-v1.6-mistral-7b-hf
This in particular is trending
I'm not sure if this is the correct…
-
Is the vision_tower you use in the code "clip-vit-large-patch14-336"? I can't load vision_tower when I use the "Gamma-MoD-llava-hr-7b-0.34" you provided. Even if I download "clip-vit-large-patch14-336…
-
### 🐛 Describe the bug
Category | Name | Inductor vs. Eager [XPU] | Inductor vs. Eager [CUDA] | XPU vs. CUDA [Eager] | XPU vs. CUDA [Inductor]
-- | -- | -- | -- | -- | --
huggingface_amp_fp16_tra…
-
For example, what transformers version should I use for llava-v1.6-mistral-7b-hf? I have tried many versions, but they all have different errors, such as
1: 4.39.2
File "/weilai/codes/lmms-eval…