-
I am currently attempting to port a llama-like model architecture from pure pytorch to TransformerEngine's pytorch classes.
However, I have been unable to obtain identical results in certain cases.…
-
# 🐛 Bug
# ❓ Questions and Help
My job can successfully run with flash attn1.0 and fail with 2.0. After I reduce the batch size to a much smaller number, it can run with 2.0. Thus, I think there …
-
Using the install script on Windows, it looks like there is a version mismatch with the dependencies it sets up in the environment, when I start up the software straight after using the install script…
-
hello
i followed your instructions for installation however when i got to run `python gradio_seesr.py` i get the following:
`WARNING[XFORMERS]: xFormers can't load C++/CUDA extensions. xFormers w…
-
### News
- 정부부처/지자체 모두 ChatGPT & 초거대AI 열공 모드: 앞으로 더많이 할 듯..
- [교육부](https://n.news.naver.com/mnews/article/079/0003736942?sid=102), [과기정통부](https://n.news.naver.com/mnews/article/421/0006645964?si…
-
### What happened?
I just received this error on update on 11-06-24
### What did you expect would happen?
It to load without any errors
### Relevant log output
```shell
The installed version of b…
-
WARNING[XFORMERS]: xFormers can't load C++/CUDA extensions. xFormers was built for:
PyTorch 2.0.1+cu118 with CUDA 1108 (you have 2.0.1+cpu)
Python 3.11.5 (you have 3.11.7)
Please reinsta…
-
**Summary**
I'm hitting a NaN loss issue when I use the TransformerLayer in place of a pytorch transformer layer I wrote.
**Details**
I'm using the nvcr.io/nvidia/pytorch:24.04-py3 docker cont…
-
### Is there an existing issue for this?
- [X] I have searched the existing issues and checked the recent builds/commits
### What happened?
xformers is installed and available in my conda env yet n…
-
# ❓ Questions and Help
memory_efficient_attention fw produce inconsistent results
not sure what was going on? incorrect built? some specific versions combinations?
for some combinations:
xfo…