NotImplementedError: No operator found for memory_efficient_attention_forward with inputs:
query : shape=(2, 2688, 1, 512) (torch.float16)
key : shape=(2, 2688, 1, 512) (torch.float16)
value : shape=(2, 2688, 1, 512) (torch.float16)
attn_bias : <class 'NoneType'>
p : 0.0
decoderF is not supported because:
max(query.shape[-1] != value.shape[-1]) > 128
xFormers wasn't build with CUDA support
requires device with capability > (7, 0) but your GPU has capability (6, 1) (too old)
attn_bias type is <class 'NoneType'>
operator wasn't built - see python -m xformers.info for more info
flshattF@0.0.0 is not supported because:
max(query.shape[-1] != value.shape[-1]) > 256
xFormers wasn't build with CUDA support
requires device with capability > (8, 0) but your GPU has capability (6, 1) (too old)
operator wasn't built - see python -m xformers.info for more info
cutlassF is not supported because:
xFormers wasn't build with CUDA support
operator wasn't built - see python -m xformers.info for more info
smallkF is not supported because:
max(query.shape[-1] != value.shape[-1]) > 32
xFormers wasn't build with CUDA support
dtype=torch.float16 (supported: {torch.float32})
operator wasn't built - see python -m xformers.info for more info
unsupported embed per head: 512
NotImplementedError: No operator found for
memory_efficient_attention_forward
with inputs: query : shape=(2, 2688, 1, 512) (torch.float16) key : shape=(2, 2688, 1, 512) (torch.float16) value : shape=(2, 2688, 1, 512) (torch.float16) attn_bias : <class 'NoneType'> p : 0.0decoderF
is not supported because: max(query.shape[-1] != value.shape[-1]) > 128 xFormers wasn't build with CUDA support requires device with capability > (7, 0) but your GPU has capability (6, 1) (too old) attn_bias type is <class 'NoneType'> operator wasn't built - seepython -m xformers.info
for more infoflshattF@0.0.0
is not supported because: max(query.shape[-1] != value.shape[-1]) > 256 xFormers wasn't build with CUDA support requires device with capability > (8, 0) but your GPU has capability (6, 1) (too old) operator wasn't built - seepython -m xformers.info
for more infocutlassF
is not supported because: xFormers wasn't build with CUDA support operator wasn't built - seepython -m xformers.info
for more infosmallkF
is not supported because: max(query.shape[-1] != value.shape[-1]) > 32 xFormers wasn't build with CUDA support dtype=torch.float16 (supported: {torch.float32}) operator wasn't built - seepython -m xformers.info
for more info unsupported embed per head: 512