-
Imported from SourceForge on 2024-07-05 22:39:48
Created by **[charpent](https://sourceforge.net/u/charpent/)** on 2016-11-12 10:49:06
Original: https://sourceforge.net/p/maxima/bugs/3239
---
Maxima…
rtoy updated
2 months ago
-
Hi, I'm trying to fine-tune the Llama3.1 8b model but after fine-tuning it uploading it to HF, and when trying to run it using vLLM I get this error "KeyError: 'base_model.model.model.layers.0.mlp.dow…
-
It's a requirement of the SPIR-V specification (https://registry.khronos.org/SPIR-V/specs/unified1/SPIRV.html#OpPhi) that "There must be exactly one Parent i for each parent block of the current block…
-
It looks like that our vectorisation strategy is to have some in-loop reduction/dependencies for a simple reduction like this:
```
for (int i = 0; i < N; i++) {
sum += a[i];
```
Because w…
-
- [x] MiniCPM-Llama3-V-2_5
- [x] Florence 2
- [x] Phi-3-vision
- [x] Bunny
- [x] Dolphi-vision-72b
- [x] Llava Next
- [ ] Idefics 3
- [ ] Llava Interleave
- [ ] Llava onevision
- [ ] internlm…
-
## Describe the bug
Disabling KV cache on the `mistralrs-server` bin via `--no-kv-cache` (as a measure to slightly reduce VRAM at the expence of the compute) leads to the garbage output from the mo…
-
### System Info
Hello TensorRT-LLM team! 👋 I'm facing an issue where the inference output does not contain the expected "Singapore" text. Below are the details of my setup and steps to reproduce the …
-
command is
`
python -m fastchat.serve.model_worker --model-names Triplex --model-path /data/cache/Triplex --controller-address http://127.0.0.1:21001 --worker-address http://127.0.0.1:8082 --host 0…
-
I am interested in calculating the GradICON and ICON loss for a network I am working on, and it does not follow the typical registration workflow with two input images and predicting the deformation f…
-
```
➜ ~ gaianet start
[+] Checking the config.json file ...
You already have a private key.
[+] Starting Qdrant instance ...
Qdrant instance started with pid: 5538
[+] Starting Llam…