-
### Expected Behavior
Troubleshoot the flux-fp8-dev black graph
### Actual Behavior
Running flux-fp8-dev e4m3fn on 4090 using Vincennes graphs will always be black, it still appears after disabling…
-
I found this issue when working with the lmms-lab/llava-onevision-qwen2-7b-ov model and qwen2vl.(the transformers library is the latest version.)
### Code
```python
import json
import argparse…
-
I can do the following to search for papers: `curl 'https://huggingface.co/api/papers/search?q=attention'`
And I get this:
>[{"id":"2409.07146","title":"Gated Slot Attention for Efficient Linear…
-
Here is the development roadmap for 2024 Q4. Contributions and feedback are welcome ([**Join Bi-weekly Development Meeting**](https://t.co/4BFjCLnVHq)). Previous 2024 Q3 roadmap can be found in #634.
…
-
I’m giving up. The files are writable and readable, but the error still appears. Nothing seems to fix it.
---------------------------------------------------------------------------
PermissionEr…
-
I am trying to use the llama3-llava-next-8b model, and I replaced --model-path with the local path of llama3-llava-next-8b that I downloaded.
When I run python -m llava.serve.model_worker --host 0.0…
-
Hi, I'm new to NLP, and I am currently trying to finetune jina for text similarity comparison.
I construct a dataset with columns `sentence1`, `sentence2` and `score`. And I can easily train the mod…
-
**Describe the bug**
Query_input's shape is [batch, pos, n_heads, d_model], and the purpose of the code where the error occurred is to reshape query_input to [batch, pos, n_heads, d_head].
I found t…
-
Hello!
The `main` (`a441a3f`) branch of the AQLM repository does not support `flash attention 2`. The error occurs because QuantizedWeight does not have a weight attribute ([closed issue #31](https…
-
When will support for batch size > 1 be available, or where should I make modifications to enable this feature?