-
From my own experience in text generation models, I found out that quantizing the output and embed tensors to f16 and the other tensors to q6_k (or q5_k) gives smaller files and better results that qu…
-
I think 1 is guaranteed to be printed out because of join(), but does join() also guarantee that 1 will be printed out before the other numbers are printed out? If yes/ no then why
![Picture1](http…
-
# Summary
I just wanted to clarify my answers for this question
## Suggested Answer (if any):
Im not sure if this is correct but my first guess for part A is 98 while for part B it is 97. However…
-
I made the `touches` array and touch functions work in the latest version of q5.js
https://github.com/quinton-ashley/q5.js
-
Oogabooga text-generation-web-ui engine used for inference (prompts directly input into the oogabooga ui produce normal results but chat-ui is doing something weird as below), Mongodb setup
_**Prom…
-
### Summary
Below is my command to start the service:
wasmedge --dir .:. --nn-preload default:GGML:CPU:qwen1_5-7b-chat-q5_k_m.gguf **llama-api-server.wasm** -p chatml -r '' --log-prompts
wa…
-
[voir retour vidéo]()
1. Q1
- [ ] => types => numérique => char vs int ?
https://github.com/PBYetml/POBJ_TestTheo_RepStudent/blob/c466e4ae6efb1e2b5a06f4549303ac3490a83e59/ASN/Test02_ASN/Test2S3…
-
./bin/gaianet init
[+] Checking the config.json file ...
[+] Downloading Meta-Llama-3-8B-Instruct-Q5_K_M.gguf ...
* Using the cached Meta-Llama-3-8B-Instruct-Q5_K_M.gguf in /Users/XXX/gaianet…
-
I am getting this error:
```
llama.cpp: loading model from /Documents/Proj/delta/llama-2-7b-chat/ggml-model-q5_1.bin
error loading model: unrecognized tensor type 14
llama_init_from_file: failed…
-
option 1 should be correct, but given answer is 2nd.
tested [here](https://rextester.com/ZJDHG66663).
also code has typo inner call should have fun4 instead of fun1.