-
### The model to consider.
https://huggingface.co/dunzhang/stella_en_1.5B_v5
last_hidden_state = model(**input_data)[0]
in __init__ model:
vector_linear = torch.nn.Linear(in_features=model.conf…
-
Hi. I am doing a project and I would like to get mid soundtrack files from Pokemon Red. You don't have such a converter yet. I would be very happy if you make one, thanks for your attention
baalp updated
31 minutes ago
-
I am using [this pytorch provided script](https://github.com/pytorch/pytorch/blob/main/benchmarks/transformer/score_mod.py) to benchmark flex attention with eager and got the attached results ([defaul…
-
Can this project help for you? https://github.com/philipturner/metal-flash-attention
So far, metal-flash-attention can indeed provide the fastest generation speed for stable diffusion on MacOS.
-
哈喽作者,感谢您的分享!在浏览您的论文时注意到attention map的可视化,想请教下类似这类注意力图的思路或者参考代码。感谢您的回复!
![d868f961d8b4fd09e2a70aca4ee9951](https://github.com/user-attachments/assets/a2f578f3-0628-48d9-87ed-6bc6801d694f)
-
The feature section lacks interactivity and visual appeal, making it less engaging for visitors, which may hinder the website's ability to capture attention.
Description of the solution I'd like
…
-
# ❓ Questions and Help
Hi All,
Debian 13
python3.10.12 venv
PyTorch2.4.1_rocm
When I try and compile xformers against Pytorch2.4.1_rocm I am ending up with the common "no file found at /th…
-
Nice paper on making LLM fully attention-based. However, I noticed that the largest model discussed in the paper is a 1.5B model.
I wonder if the pattention layer is difficult to tensor parallelize…
-
### System Info
```Shell
- `Accelerate` version: 0.34.2
- Platform: Linux-5.4.0-45-generic-x86_64-with-glibc2.31
- `accelerate` bash location: /home/gradevski/miniconda3/envs/summary_explainer_p…
-
Hi, I have paid attention to your project and I think it is very good, but I still don't know what the output is. Could you explain it to me roughly? Can it render guassiansplat