-
### 🐛 Describe the bug
When initializing a Transformer like this
```
nn.Transformer(hidden_dim * 2, 4, batch_first=True)
```
and then calling it like this
```
attention_mask = nn.Transformer.…
-
### 🐛 Describe the bug
``` Python
from functools import lru_cache
from torch.nn.attention.flex_attention import flex_attention, create_block_mask
import torch
torch._dynamo.config.cache_s…
-
### System Info
- `transformers` version: 4.46.2
- Platform: Linux-6.1.85+-x86_64-with-glibc2.35
- Python version: 3.10.12
- Huggingface_hub version: 0.24.7
- Safetensors version: 0.4.5
- Accele…
-
**Issue Title**: RuntimeError: cutlassF: No kernel found to launch in `sample_t2v.py`
**Description**:
I encountered an error when running the code in `opensora/sample/sample_t2v.py`. The specif…
-
**LocalAI version:**
OK:
- `local-ai-avx2-Linux-x86_64-1.40.0`
- `local-ai-avx2-Linux-x86_64-2.0.0`
- `local-ai-avx2-Linux-x86_64-2.8.0`
- `local-ai-avx2-Linux-x86_64-2.8.2`
- `local…
-
### Describe the issue
The error is only raised after 1.18. I tried 1.17.3, and it works fine.
onnxruntime.capi.onnxruntime_pybind11_state.InvalidArgument: [ONNXRuntimeError] : 2 : INVALID_ARGUMEN…
-
```python
import time
import torch
from torch.nn import functional as F
from flash_attn import flash_attn_func
from einops import rearrange
import math
def standard_attention(query_layer, key…
-
### 🐛 Describe the bug
when I set the `dropout_p=0.0`, the result is different. But `dropout_p=-1`, the result is same. Maybe the op scaled_dot_product_attention has some bug. Please fix it, thank…
-
### Describe your issue
When I have a large diagram, I often create views with certain nodes hidden, so that certain aspects of the diagram can be focused on. But if a connecting node is hidden, th…
-
I want to fine-tune a model using unsloth. Every thing works fine on colab but on my system I got the following:
{
"name": "NotImplementedError",
"message": "No operator found for `memory_efficie…