-
out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd(
RuntimeError: CUDA error: device-side assert triggered
CUDA kernel errors might be asynchronously reported at some ot…
-
I'm trying to run the following code in kaggle with **GPUP100**
`!bash /kaggle/working/Phi3-Vision-Finetune/scripts/finetune_lora_vision.sh`
### complete error
`[2024-09-14 09:33:24,960] [INFO] …
-
Now LLAMA 3.1 is out, but sadly it is not loadable with current text-generation-webui. I tried to update transformers lib which makes the model loadable, but I further get an error when trying to use …
-
### System Info
When attempting to add localdocs plugin, nothing shows up when it attempts to download from: GPT4ALL.IO/MODELS/MODELS2.JSON
it looks like broken link issue ?
### Information
- […
-
> Please provide us with the following information:
> ---------------------------------------------------------------
### This issue is for a: (mark with an `x`)
```
- [x] bug report -> please…
-
The fine-tuning code runs when I replace the base model with LLAMA-2.
I am aware that LLAMA and LLAMA-2 share the same configuration files and other associated components.
However, I would sti…
-
Package Version:
AutoAWQ: 0.2.5+cu118
torch: 2.3.1+cu118
transformers: 4.43.3
I was try to quantize my finetuned llama3.1 405b (bf16) model to 4 bit using autoawq following the insturction in t…
-
Hey spuun @spuuntries , I've been using your model and it seems to work great, but I've tried to add history/context to the model by adding a ```### History:``` section to the prompt with langchain.
…
ylhan updated
5 months ago
-
-
- [ ] [README.md · defog/sqlcoder-7b-2 at main](https://huggingface.co/defog/sqlcoder-7b-2/blob/main/README.md?code=true)
# README.md · defog/sqlcoder-7b-2 at main
**DESCRIPTION:**
```yaml
license:…