Closed hubertwang closed 2 weeks ago
Ah, sorry, I just noticed that, in the example of README, there's a ==n_ctx== parameter But it is absent in the example code. When I added it to initLlama, it start generating meaningful response.
Not sure why android don't need this, but anyway I found the solution.
initLlama({
model: file.uri,
use_mlock: true,
n_ctx: 2048, // added this for iOS
n_gpu_layers: Platform.OS === 'ios' ? 0 : 0, // > 0: enable GPU
// embedding: true,
})
Hi everyone,
I tried the example app, and runs it on my devices. On android devices, it works perfectly without any issue. But on iPhone, it generate irrelevant outputs.
(Model: gemma-2-2b-it-Q8-0.gguf) For example, when I ask: "What are the three laws of robotics?" It answered:
Looks like the input and output are not related. Anyone has the same issue?