EricLBuehler / xlora

X-LoRA: Mixture of LoRA Experts
Apache License 2.0
178 stars 8 forks source link

seq_len is always 1 #32

Closed Zehrooo closed 4 months ago

Zehrooo commented 4 months ago

Hello, I found that during model inference, when I print "model.internal_xlora_scalings.shape", the result is consistently [1, 1, 28, 7]. It is supposed to be [batch_size, seq_len, n_layers, n_classes], but I don't understand why seq_len is always 1.

model:ChatGLM3-6b-base peft==0.10.0 transformers==4.37.2