-
hi getting this error while using the package , i might be using it wrong, can you help me with
``` const llmInference = useLlmInference({
storageType: 'file',
modelPath: './gemma-2b-it…
-
# Model name
Google Gemma family (7B, 2B, 7B-instruct, 2B-instruct)
# Parameters
Not that I'm aware
# Source
Models are available via huggingface (`transformers`):
7B: https://huggingface.co…
-
### Describe the issue
If I create a model in Autogen studio that points to the LM studio endpoint then add the model to an agent, then a workflow etc, when I run the workflow it will terminate after…
-
Thank you for your excellent work. I'd like to know how to reproduce jbloom/Gemma-2b-Residual-Stream-SAEs
Can you open source the configuration file for training it?
-
I keep getting this error when I run with
`CUDA_LAUNCH_BLOCKING=1; tune run --nproc_per_node 4 lora_finetune_distributed --config scripts/2B_lora.yaml`
any thoughts what I might be doing wrong.…
-
First, thanks for putting this project together!
I modified `examples/basic/index.html` to use a more capable model: `https://huggingface.co/lmstudio-ai/gemma-2b-it-GGUF/resolve/main/gemma-2b-it-q4…
-
While generating any text with a specified value of max_length, the generated text keeps repeating several times until the output spans the value of max_length. An example of the above is using the fo…
-
I'm encountering a RuntimeError when attempting to save checkpoints during fine-tuning the "unsloth/gemma-2b-it-bnb-4bit" model. Below is a breakdown of my setup and the error encountered.
Model:…
-
**Issue: Model Error when Setting max_seq_length > 8192**
**Description:**
The `unsloth/codegemma-2b-bnb-4bit` model throws an error when attempting to set `max_seq_length` greater than 8192.
…
-
When I run `Ollama run gemma:2b` inside docker, it will download blobs folder and manifests folder in `/root/.ollama/models`.
In private enviroment without interner, I copy `blobs` & `manifests` fo…