-
Now, i quantize llama3-8b model using QAT. When I tried model inference, I encountered the following error.
Logs:
```bash
une run generate --config llama3_generation_config.yaml
2024-10-09:08:06:2…
-
Hi team,
I would like to fine tune Llama2 model with my domain data and eventually convert it as a chat model. so, I can directly ask questions related to my domain data and get right response from…
-
Hello everyone.
In gemma-7b looks share embedding layer's weights to lm_head. So I try to train my own lm_head that updated independently.
First, Load model
```python
model = AutoModelForCausa…
-
Checking to see if this repo works for the new L3 models. Running this script:
```
export CUDA_VISIBLE_DEVICES=0,1 # or e.g. 0,1,2,3
export MODEL_PATH=/home/catid/models/Meta-Llama-3-8B-Instru…
catid updated
7 months ago
-
Hi!
First of all thanks for the nice repo!
I tried already many different proposed solutions from here:
https://github.com/oobabooga/text-generation-webui/issues/5885
but I always get eith…
-
### System Info
- peft 0.12.0
- `transformers` version: 4.43.3
- Platform: Linux-5.15.0-113-generic-x86_64-with-glibc2.35
- Python version: 3.12.4
- Huggingface_hub version: 0.24.2
- Safetensors…
-
For ReadMe step under Usage: "Activate the plugin through the editor's toolbar."
In Plugins, I can see 'blupy', and it is ticked meaning enabled. I created a new Blueprint with parent of Object. In…
-
### What is the issue?
I recently upgraded my GPU from an RX 6800 to an RX 7800 XT. Both GPUs are supported by ollama and everything ran fine on the 6800. If I run the same docker container on the …
-
### System Info
pip list |grep -i -E 'cuda|torch'
nvidia-cuda-cupti-cu12 12.1.105
nvidia-cuda-nvrtc-cu12 12.1.105
nvidia-cuda-runtime-cu12 12.1.105
torch 2.3.0
GPU i…
winca updated
6 months ago
-
Hello! When I was using your code for training, I found that the metrics for refcoco are always very low, even when reproducing your DINOv2 + SigLIP 384px (Naive Resize) configuration. During finetuni…