Closed fxmarty closed 10 months ago
cc @sabreshao @howiejayz @fsx950223 what do you think?
Unrelated - I was wondering if you were open to allow issues in this repo? I encountered a few that I think could be nice to report (at least for other users using this repo).
@fxmarty we plan to add an option to resolve docker build. @howiejayz will do that.
Hi @fxmarty, can you close this PR and move any of your request to issues? I will go through them including this one.
Hi @howiejayz happy to do so - however I can't see an issues tab in the repo:
Hi @howiejayz happy to do so - however I can't see an issues tab in the repo:
Hi @fxmarty, could you try the latest build_and_run.sh script for building flash-attention in Dockerfile. Also the Issue section is finally opened.
Hi @howiejayz, thank you I indeed noticed the available GPU_ARCHS
variable, that's working fine now.
Thank you for opening the issue section!
Hi,
Hard-coding
--offload-arch=native
make the build of RoCm flash attention fail in docker build (as I guess GPUs are not accessible during build)Moreover, this prevents
setup.py
to obey to the variablePYTORCH_ROCM_ARCH
, which is a quite useful feature.