Open amdrenwuli opened 7 months ago
@amdrenwuli Please see this pr. In order to build FA in the dockerfile, you need to manually set --offload-arch to the type of GPUs you have. This is because the environment in the dockerfile has no access to your GPUs.
@amdrenwuli Please see this pr. In order to build FA in the dockerfile, you need to manually set --offload-arch to the type of GPUs you have. This is because the environment in the dockerfile has no access to your GPUs.
Thanks for your reply. Another question, do we not support installing csrc/rotary and csrc/layer_norm currently?
root@xcdlossgpu08:/workspace/flash-attention# pip install csrc/layer_norm/
Processing ./csrc/layer_norm
Preparing metadata (setup.py) ... error
error: subprocess-exited-with-error
× python setup.py egg_info did not run successfully.
│ exit code: 1
╰─> [13 lines of output]
Traceback (most recent call last):
File "<string>", line 2, in <module>
File "<pip-setuptools-caller>", line 34, in <module>
File "/workspace/flash-attention/csrc/layer_norm/setup.py", line 99, in <module>
raise_if_cuda_home_none("--fast_layer_norm")
File "/workspace/flash-attention/csrc/layer_norm/setup.py", line 46, in raise_if_cuda_home_none
raise RuntimeError(
RuntimeError: --fast_layer_norm was requested, but nvcc was not found. Are you sure your environment has nvcc available? If you're installing within a container from https://hub.docker.com/r/pytorch/pytorch, only images whose names contain 'devel' will provide nvcc.
torch.__version__ = 2.3.0a0+gitac0bed0
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
error: metadata-generation-failed
× Encountered error while generating package metadata.
╰─> See above for output.
note: This is an issue with the package mentioned above, not pip.
hint: See above for details.
No. I think the CK team might have some plans. But for now I can confirm we do not have those features.
@amdrenwuli Support for other modules (rotary, xentropy, layer_norm) We will track effort in that issue.
Problem Description
WORKDIR /workspace USER root
RUN pip install ninja RUN git clone -b flash_attention_for_rocm --recurse-submodules https://github.com/ROCmSoftwarePlatform/flash-attention.git RUN cd /workspace/flash-attention \ && python setup.py install RUN pip3 list
Operating System
20.04.5 LTS (Focal Fossa)
CPU
AMD EPYC 73F3 16-Core Processor
GPU
AMD Instinct MI250X
ROCm Version
ROCm 5.7.0
ROCm Component
No response
Steps to Reproduce
Shown in Problem Description
(Optional for Linux users) Output of /opt/rocm/bin/rocminfo --support
rocminfo.txt
Additional Information
No response