Vahe1994 / SpQR

Apache License 2.0
515 stars 40 forks source link

Can I save the compressed model for direct inference only? #1

Closed SparkJiao closed 1 year ago

SparkJiao commented 1 year ago

Excellent work

May I know if I could save the compressed model locally for further inference, e.g., combined with lora adapters?

I see the NotImplementedError raised so I'm not sure if there is something should be noted.

Thanks!

Vahe1994 commented 1 year ago

Hey!

Thanks for the interest in our work!

We've released only the evaluation code for now (you can use it to evaluate the compressed model's quality).

We'll add code for efficient inference soon (in ~2 weeks).

SparkJiao commented 1 year ago

Thanks for your clarification!

singingtower commented 8 months ago

Hey!

Thanks for the interest in our work!

We've released only the evaluation code for now (you can use it to evaluate the compressed model's quality).

We'll add code for efficient inference soon (in ~2 weeks).

has the inferance code released by now? failed to find it!