-
see #27
https://ai.google.dev/gemma/docs?hl=en
https://www.kaggle.com/models/google/gemma
Gemma on Vertex AI Model garden
https://console.cloud.google.com/vertex-ai/publishers/google/model-gard…
-
**LocalAI version:**
```
v1.25.0-cublas-cuda12-ffmpeg
```
**Environment, CPU architecture, OS, and Version:**
```
# uname -a
Linux localai-ix-chart-f8bbbb7c7-x6xx9 6.1.42-production+truen…
-
Trying to load some more recent Q5_K_M models using talk-llama and getting errors about tensor type 13. @ggerganov can you please update talk-llama to work with the latest llama.cpp?
Thank you!
-
Hi! I have followed every step in [Run Llama 2 on your own Mac using LLM and Homebrew](https://simonwillison.net/2023/Aug/1/llama-2-mac/), in particular:
```
pipx install llm # python 3.11
llm in…
-
### 是否已有关于该错误的issue或讨论? | Is there an existing issue / discussion for this?
- [x] 我已经搜索过已有的issues和讨论 | I have searched the existing issues / discussions
### 该问题是否在FAQ中有解答? | Is there an existing…
Lyzin updated
2 months ago
-
Not sure if this RNN counts as a LLM, but if so would be nice to have it, let me know what needs to be done with packaging.
https://www.rwkv.com/
-
Hi all.
I just got a Microsoft laptop7, the AIPC, with Snapdragon X Elite, NPU, Adreno GPU. It is an ARM based system.
But I found that NPU is not running when using Ollama.
Would it be suppo…
-
### What happened?
cmd: docker run --rm -it --gpus all ghcr.nju.edu.cn/ggerganov/llama.cpp:full-cuda --version
output:
```
docker: Error response from daemon: failed to create task for container:…
-
Hey, thank you so much for the great model and this repo!
Would you be willing to add support for this chat format to llama-cpp-python, so that we can use function calling (and JSON mode) with thei…
-
I’ve discovered a performance gap between the Neural Speed Matmul operator and the Llama.cpp operator in the Neural-Speed repository. This issue was identified while running a benchmark with the ONNXR…