[NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support Llama-3/3.1, Llama-2, LLaMA, BLOOM, Vicuna, Baichuan, TinyLlama, etc.
Hi! Is it possible to save the model and create custom configuration files so that we can push to hugging face and load? Also, can PEFT be used directly from transformers library, or should we load peft.from_pretrained from this LLM-Pruner repo?
Hi! Is it possible to save the model and create custom configuration files so that we can push to hugging face and load? Also, can PEFT be used directly from transformers library, or should we load peft.from_pretrained from this LLM-Pruner repo?