-
### Describe your use-case.
Flux has layers named single_transformer_blocks.* and transformer_blocks.*.
If I want to train only the **transformer_blocks.*** layers but exclude **single_transformer…
-
All of the following are linear transformers:
- `center()`
- `standardize()`
- `slide()`
- `reverse()`
- `rescale()`
- `normalize()`
So hypothetically we should be able to simplify their co…
-
Running this command
```
CUDA_VISIBLE_DEVICES=0 python3 model/llama.py ShiftAddLLM/Llama-2-70b-wbits2-acc
```
I'm seeing this error
```
CUDA extension not installed.
Loading checkpoin…
-
Thanks for your nice contribution!!
When I try to replace the Transformer block in a model with VSSEncoder(The Transformer includes factorized self-attention for its linear complexity as done in…
-
https://arxiv.org/abs/2006.16236
-
A simple and elegant work and it seems to be the state-of-the-art graph transformer for node classification.
I notice that the largest dataset used in your paper is ogbn-products with about 2 milli…
-
When opening the URL (http://0.0.0.0:7860) I get the "can't reach this page" message. I don't get any errors while loading, apart from the "No module named 'triton'" one, which I assume is normal on …
-
**Is your feature request related to a problem? Please describe.**
Would be great to be able to load a LoRA to a model compiled with `torch.compile`
**Describe the solution you'd like.**
Do `load…
-
### Describe the bug
I trained a lora with simpletuner using ai-toolkit preset (I used all+ffs and others and it doesnt train correctly on hard concepts).
And Now I have this issue when loading the …
-
Thanks for your great work!
https://github.com/liangyanshuo/InfLoRA/blob/2c774547d48c40fe5bb8c4a393f7b370e1664148/models/vit_inflora.py#L237
I have a question regarding this part of the code. In…