Open hrQAQ opened 1 year ago
changelog from transformers Release 4.35.0
Attention mask refactor
We refactored the attention mask logic for major models in transformers. For instance, we removed padding_mask argument which was ambiguous for some users
Remove ambiguous padding_mask and instead use a 2D->4D Attn Mask Mapper by @patrickvonplaten in https://github.com/huggingface/transformers/pull/26792 [Attention Mask] Refactor all encoder-decoder attention mask by @patrickvonplaten in https://github.com/huggingface/transformers/pull/27086
By running the following commands, my windows server is successfully started.
pip uninstall transformers
pip install transformers==4.34.0
So, I suggest modifying the setup.cfg to make minimal changes.
Thank you for reporting this. I'm fixing transformers version for now. In the meantime, we're working on fixing the problem in upstream
I'm following the tutorial Run-Petals-server-on-Windows to start up a server on my own PC. Upon running
python -m petals.cli.run_server petals-team/StableBeluga2
, I encountered the following error:After STFW, I found that the root cause of this error report mab be related to the upstream refactored the attention_mask module and the related commit page is here.
I propose there are two possible solutions to this issue. The first one is to specify the download of a previous version of the 'transformers' library when installing dependencies. The second solution is to adapt to the new attention mask implementation(needs some modification of petals/models/llama/block.py).