-
## System:
MacBook: Apple M1 8 GB MacOS Sonoma v 14.1.1
wasmedge version: 0.13.5
## Issue Description:
I am encountering an error when attempting to run a llama2 model using WasmEdge. The comman…
-
Hello I wish to clear the models knowledge for every loop iteration without tracking previous inputs or outputs. Since the context object is a way to pass information between different parts of the mo…
-
Just after i press load model i get error loading file and empty modal path withou even opening file picker , storage permission are granted
![Screenshot_20231122-024934](https://github.com/MaidFound…
-
not an issue
- Install TermUX form Fdroid
- pkg upgrade
- pkg install git
- pkg install make
- pkg install clang
- pkg install wget
- git clone https://github.com/trholding/llama2.c…
-
python test_inference.py -m TinyLlama-1B-4.0bpw-h6-exl2 -p "Once upon a time,"
```
-- Model: TinyLlama-1B-4.0bpw-h6-exl2
-- Options: ['rope_scale 1.0', 'rope_alpha 1.0']
-- Loading model...
…
-
Hello,
I am trying to execute the llama file on arm64 linux: `Linux rock-5b 5.10.110-rockchip-rk3588 #23.02.2 SMP Fri Feb 17 23:59:20 UTC 2023 aarch64 aarch64 aarch64 GNU/Linux`
I added ape to the b…
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [x] I am running the latest code. Development is very rapid so there are no tagged versions as of…
hksk updated
11 months ago
-
Hello, tiny lama takes all my ram and has very very poor perfs' like lower than 7b models, it takes a very long time to load and is worse than most model, I don't unstand what I'm doing wrong, usually…
-
After loading `TheBloke/CodeLlama-7B-Instruct-GGUF/codellama-7b-instruct.Q5_K_M.gguf`, as soon as I send an input I get the following error and crash:
```
ggml_init_cublas: found 2 ROCm devices:
…
-
Hi to whoever is reading this 🤗
## Description
I'm having a hard time trying to reproduce and obtain https://huggingface.co/mlx-community/Mistral-7B-Instruct-v0.2-4-bit using the recently includ…