-
## 🐛 Bug
I've tried in my `OnePlus 12R` with the processor of `Snapdragon 8 Gen 2` and `8GB` of RAM.
I just Installed the `MLCChat` Apk, and downloaded the `gemma-2b-q4f16_1` model in my phone. Af…
-
Can this work with LM Studio which allows different models to be used more than Ollama?
-
### What is the issue?
hi? I'm studying fine tuning.
I learned using the "unsloth/gemma-2-2b-it" model.
I created the dataset myself and it contains less than 100 cases.
I want to use only the fin…
-
Hugginface hub login successful
Used gemma2-27b LLM to testing:
cargo run --release -- -m "google/gemma-2-27b-it" -c
Finished release [optimized] target(s) in 0.03s
Running `target/re…
-
I used [llm_inference](https://github.com/googlesamples/mediapipe/tree/main/examples/llm_inference) sample with `gemma-2b-it-cpu-int4.bin` on Pixel 8 Pro emulator.
The prefill speed seems to be in…
-
Right now the vector DB is working (#7) and we also made the ANN distance thresholds configurable (#35), but for proper RAG it'd be great to have re-ranking. Using Gemini this could mean many calls. M…
-
[ ] I have checked the [documentation](https://docs.ragas.io/) and related resources and couldn't resolve my bug.
**Describe the bug**
"ValueError: a cannot be empty unless no samples are taken"
…
-
This is a very valuable project for research. I tried to download it. The demo cpp llama1.1b w4a8 can run at 18 t/s. Although the output of the model is not what I asked, it seems that there is no pro…
-
First, thanks for putting this project together!
I modified `examples/basic/index.html` to use a more capable model: `https://huggingface.co/lmstudio-ai/gemma-2b-it-GGUF/resolve/main/gemma-2b-it-q4…
-
I'm trying to use gemma with unreal engine and I got an exception in an unexpected place:
```
gcpp::N_AVX2::Decompress(gcpp::CompressedArray const & __ptr64,unsigned __int64,float * __ptr64,unsigned…