-
@bin123apple Where do you share the pairs of before (Fortran) vs. after (C++) translation so @peihunglin can start initial manual grading of them ?
-
Hi team,
I was fine tuning an LLM with Ludwig on a **NVIDIA A 100** instance.
I get the error message - **Encounted `nan` values in tensor. Will be removed.", UserWarning)** My loss and perplexi…
-
Thanks for your effort. I have a little confusion about the process. Correct me if I'm wrong. First, we should run block_expansion.py to create our extended model. Then, we clone the repository at…
-
### System Info
**System Info:**
GPU used : A100 80 GB
Pytorch version : 2.4.1+cu124
CUDA used to build PyTorch: 12.1
### Information
- [X] The official example scripts
- [ ] My own modified s…
-
Hi, after going through the training code, it seems that the gradient is not properly backpropagated. It seems that all projector layers ```mm_projector```are called within ```torch.no_grad``` (i.e.,…
-
When I am doing finetune llama3.1 does the following error occurs, can't locate the exact error, how to fix it please?
Running environment:
```
Python 3.11.0rc1
GPU: 2xA100
CUDA Version: 12.2
…
-
I refined llama3.1 8b bnb 4bits according to your recommendations with my own train+eval dataset and saved as merged 16 bits. I now want to create an inference by loading the 16b merged model and usin…
-
### Your current environment
Collecting environment information.
PyTorch version: 2.3.0+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
0S: Cent0S Li…
-
Hi there,
I am running the DNABERT run_finetune.py as instructed by the readme file. It works well at my workstation, but when I run the same code on the server, it reports the following error:
…
-
The GUI is not working.
(gpt1) ashu@MSI:/mnt/c/Users/genco/Documents/gpt$ make run
poetry run python -m private_gpt
11:47:15.734 [INFO ] private_gpt.settings.settings_loader - Starting appli…