-
Chapter 3 explains the method to download the LLama.cpp model weights and tokens. This chapter specified 3B model. I have two question:
1- The available model weights and tokens are for models 7B and…
-
### What happened?
After installing and logging into the Quivr on Amazon Linux with default userid/password, **not able to create the first brain**. Request your help as we are stuck with this iss…
-
I've been trying to hunt down a bug. I think I've pinned down that there is some harmful interaction between `PPOTrainer` & `AutoModelForCausalLMWithValueHead` & `Gemma 2 2B`
**Edit:** In the comme…
-
I have tested Ollama on different machines yet, but no matter how many cores or RAM I have, it's only using 50% of the cores and just a very few GB of RAM.
For example now I'm running `ollama rum lla…
-
how to run nexa with locall modele?
-
### Summary
Using the following command line to install Rust TLS plugin and ggml plugin.
```
curl -sSf https://raw.githubusercontent.com/WasmEdge/WasmEdge/master/utils/install.sh | bash -s -- --p…
-
Hi, thanks for your good work, guys. i'm running ollama on ArchLinux. when I start the ollama service and get a status, it shows these logs:
> hub.com/jmorganca/ollama/server.(*Server).GenerateRout…
-
### Model ID
four-two-labs/lynx-micro
### Model type
Decoder model (e.g., GPT)
### Model languages
- [ ] Danish
- [X] Swedish
- [ ] Norwegian (Bokmål or Nynorsk)
- [ ] Icelandic
- [ ] Faroese
- […
-
The SIMD acceleration on x86_64 does not seem to be as optimized as on AArch64. Perhaps some optimization work is needed for the x86_64 platform.
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [ ] I am running the latest code. Development is very rapid so there are no tagged versions as of…