-
From https://arxiv.org/abs/2112.05682v2. I have no immediate use for this, but it looks cool and I didn't want it to go unmentioned in case some aspiring contributor to Transformers.jl is looking for …
-
Can xformers be replaced with xformers-0.0.23. post1 cp39 cp39 win amd64. whl?
I always report errors when running NotImplementedError: No operator found for `memory_efficient_attention_forward` with…
-
### Is there an existing issue for this?
- [X] I have searched the existing issues and checked the recent builds/commits
### What happened?
xformers is installed and available in my conda env yet n…
-
# 🐛 Bug
Using xformers.memory_efficient_attention with FSDP and torch.compile fails when using bfloat16, but works when using float32. It's unclear to me if this is an xformers bug, an FSDP bug, or…
-
Hello,
When trying to apply the Sine Wave example approach to a transformer based model I get the following output:
File "/usr/local/lib/python3.10/dist-packages/torch/autograd/graph.py", …
-
Liger (Linkedin GPU Efficient Runtime) Kernel is a collection of Triton kernels designed specifically for LLM training. It can effectively increase multi-GPU training throughput by 20% and reduces mem…
-
Hi! Thank you so much for this.
I'm trying to run this on Google Colab but I'm always running into the "CUDA out of memory" error.
I've tried adding:
```
+ --enable_xformers_memory_efficient_…
guivr updated
8 months ago
-
I'm a beginner to try unsloth. I run the free notebook [Llama 3 (8B)](https://colab.research.google.com/drive/1bX4BsjLcdNJnoAf7lGXmWOgaY8yekg8p?usp=sharing#scrollTo=yqxqAZ7KJ4oL), and then got the fol…
-
### Python Version
```shell
Python 3.10.12 (main, Mar 22 2024, 16:50:05) [GCC 11.4.0]
```
### Pip Freeze
```shell
absl-py==2.1.0
annotated-types==0.7.0
attrs==23.2.0
docstring_parser=…
-
i have this error
can anybody help me ?
Error occurred when executing DynamiCrafterInterp Simple:
No operator found for memory_efficient_attention_forward with inputs:
query : shape=(80, 2560,…