Closed ChristianWeyer closed 1 year ago
I'm currently disabling falcon on MPS since I would just run out of memory to try even run the model on Mac
Not sure if this is valid any more. I have since tested a lot with pytorch on MPS, and it is often slower. Will probably investigate mlc vs. gguf for this.
Feature request
I want to use OpenLLM with available models to run on Apple M1/M2 processors (GPU support) through MPS.
Today:
Motivation
No response
Other
No response