KIVI/models/llama_kivi.py", line 25, in __init__
self.attention_dropout = config.attention_dropout
File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 262, in __getattribute__
return super().__getattribute__(key)
AttributeError: 'LlamaConfig' object has no attribute 'attention_dropout'