-
i got problem in conda:
`Traceback (most recent call last):
File "C:\Users\always sunsine\.conda\envs\barber\lib\site-packages\IPython\core\interactiveshell.py", line 3526, in run_code
exec…
-
### 🐛 Describe the bug
inductor/test_move_constructors_to_cuda.py::TestMoveConstructorsToCuda::test_multi_gpu FAILED [1.4059s] [ 14%]
==================================== RERUNS ====================…
-
Hello, I'am study fused_multi_head_attention example in CUTLASS.
In CUTLASS 3.5.1 README.md, it said flash attention 2 kernel is in CUTLASS.
But in fused_multi_head attention, it is based on Meta/xFor…
-
# Quantize the model
model_prepared = tq.prepare(model_fused)
model_quantized = tq.convert(model_prepared)
# Define the quantization configuration
quant_config = tq.get_default_qconfig('fbge…
-
both [fused-attention](https://triton-lang.org/main/getting-started/tutorials/06-fused-attention.html#sphx-glr-getting-started-tutorials-06-fused-attention-py) and [flash-attn-og](https://github.com/D…
-
Hi
Could you help me regarding this output:
python single_case_prediction.py --conf P2Y14_example/hparams.yaml --ckpt_path ckpts/with_ligand.ckpt --input_fn P2Y14_example/P2Y14_AF.pdb,P2Y14_ex…
-
**What is your question?**
when I use sm89
```
int run_attention(Options& options) {
using Attention = AttentionKernel<
cutlass::half_t, // scalar_t
cutlass::arch::Sm89, // ArchT…
-
Hello author, thank you for your contribution to the pansharpening deep learning toolbox. When I was running run_test_manshharpening. py, the saved test result (i.e. the fused image file) was empty. H…
-
def fuse_conv(conv, norm):
fused_conv = torch.nn.Conv2d(conv.in_channels,
conv.out_channels,
kernel_size=conv.kernel_size,
…
T-wow updated
1 month ago
-
Hi! Thank you for your amazing work!
I'm having some trouble on comparing the fused swiglu kernel with its plain pytorch version.
I checked the wall clock time with code below, and it gives me l…