Closed helleuch closed 10 months ago
Did you use trust_remote_code=True
? I don't have this anywhere in the documentation right now, but all models should be loaded without trust_remote_code=True
.
Thank you, I will try that.
That solved the issue :)
Hello,
When using attention sink with
falcon-7b-instruct
, I get the following error:TypeError: FalconForCausalLM.__init__() got an unexpected keyword argument 'attention_sink_size'
. I added theattention_sink_size
andattention_sink_window_size
to theAutoModelForCausalLM.from_pretrained
in my code.But in the provided example, it says it is compatible with it. Am I overlooking something ?
Edit: I tried it with mistal-7b and did not get the error.