Open fabianlim opened 1 month ago
we found out there is really no need to upper bound this torch dependency, as for us we are getting stuck only beause of this commit https://github.com/pytorch/pytorch/pull/121635.
nvidia-nccl-cu12==2.19.3
~ Update: This is due to a NCCL_BUFFSIZE
wrong setting. Also for transformers, we just have to be weary of the sliding window mask issue of SPDA, and keep track of it to see when it will be fixed.
FMS has fixed the TRL issue https://github.com/foundation-model-stack/fms-hf-tuning/pull/213
I think we need a lower limit on the bitsandbytes
version that supports quant_storage
. I have encourted that 0.41
didnt work, but 0.43
is ok
Currently the
torch
dependency inframework
is upper bounded as"< 2.3"
, however inaccelerate
versions has problems supporting torch2.2
. The latestnumpy
versions (>=2.0) also has incompatibilities with the current torch version and is bounded here in #42. Hence, we should consider releasing the upper bound soon.Also can consider releasing the upper limit on
transformers
andaccelerate