-
Whenever I try to load it, it crashes Chrome.
This is on a Pixel 6a with 6Gb of RAM.
- Context is set to 1K.
- 16 bit WebGPU is available.
- Using latest version of WebLLM from CDN.
To make…
-
Hi, thanks for providing a great examples. It is amazing indeed.
While I am deploying the test app for LLM chat, I tried two models which are Gemma and phi-2.
For Gemma, it worked perfectly. But f…
-
### Description of the bug:
I tried running the example.py script given for quantization example, but for Llama. Wherever the reference to Gemma was made, I made appropriate references to Llama. The…
-
### Context
This task regards enabling tests for **gemma-7b-it**. You can find more details under openvino_notebooks [LLM chatbot README.md](https://github.com/openvinotoolkit/openvino_notebooks/tree…
-
Just tried the new Gemma-2 9B model with a context length of around 10K Tokens. First run works fine, second run gives me the following error:
```
File "[...]exllamav2/exllamav2/generator/base.p…
-
### Model description
bge-reranker-v2.5-gemma-lightweight 's performance is better bge-m3 :)
Please support model.
### Open source status
- [ ] The model implementation is available
- [X] The …
-
Author: @vqd8a
-
### System Info
- CPU Architecture: x86_64
- CPU Type: AMD Epyc 9654
- GPU Type: Nvidia H100
- Nvidia Docker Container: nvcr.io/nvidia/nemo:24.01.gemma
### Who can help?
@Tracin
### Information…
-
Hello.
I downloaded AI Playground v1.22.1 for desktop GPUs which has a built-in "LLM picker" but unfortunately the dGPU version does not provide Mistral model- as mentioned in the release notes (it h…
-
The @PavlidisLab/curation team compiled a list of single-cell experiments that need special treatment on import. We can convert those into test cases.
## Bulk/single-cell mix
Contains a mixture …