Open gayuoptisol opened 6 months ago
Hi @gayuoptisol,
You can set the device_map
to “cpu” as follow,
llm_lingua = PromptCompressor(device_map="cpu")
hi :-) is there a rule of thumb - how much GPU memory is required depending upon model/processed content, and how does this translate to required number of CPUs and their memory?
Is there any method to run LLMLingua in linux cpu machine. I am trying to load this using: from llmlingua import PromptCompressor llm_lingua = PromptCompressor(device_map="mps")
but it taking so much amount to load.