Closed madroidmaq closed 1 week ago
Me, too. Device: Pixel 7 Pro Android14 Model: gemma-2b-it-cpu-int4.bin / gemma-1.1-2b-it-cpu-int4.bin
However, the gemma-2b-it-gpu-int4.bin works
I hope you took latest changes of llminference from the examples.
Can you measure the memory usage during the inference call? We are working on expanding our hardware support, but we are currently mostly focused on bringing inference to recent set of Android phones that provide more memory. If you are targeting Pixel 4, I would suggest using a smaller LLM such as Falcon 1B or the even smaller models outlined here: https://github.com/google-ai-edge/ai-edge-torch/tree/main/ai_edge_torch/generative/examples/smollm
Unable to successfully perform inference tasks on Google Pixel 4 device, the error message is as follows: