-
Hi Will,
Could you please confirm the code and weights used on your end?
Cheers,
Victor
-
![image](https://github.com/user-attachments/assets/cf62a87d-cf3d-4f15-a40c-17bb87bb419d)
```
parser.add_argument(
"--total-token",
type=int,
default=60,
help=…
-
Currently, TensorRT-LLM requires that LoRA weights dtype match the base model dtype. The check is here:
https://github.com/NVIDIA/TensorRT-LLM/blob/9dbc5b38baba399c5517685ecc5b66f57a177a4c/cpp/tensor…
-
**I try to get clipcap_coco_weights.pt by:**
> gdown 1IdaBtMSvtyzF0ByVaBHtvM0JYSXRExRX -O ${PT_MODEL_DIR}/clipcap_coco_weights.pt
**but encounter:**
> Access denied with the following error:
…
-
Dear developer, I am interested in your work, may I ask you if you could uplaod your trained weights?
-
### Question
Hi, Thanks for the awesome repo. I am trying to finetune your model on a custom dataset. My GPU memory is not enough to finetune relation_detr_focalnet_large_lrf_fl4_1200_2000.py version…
-
### Your current environment
```text
Collecting environment information...
PyTorch version: 2.2.1+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
…
-
When settting `os.environ["WANDB_LOG_MODEL"] = "end"` prior to the training loop and specify `report_to='wandb'` in `TrainingArguments`, I receive the following error:
```
Loading best SentenceTra…
-
Hello,
I followed the ReadMe, creating a conda environment, activating it and running the demo with hero_model and vdr dataset according to the section "Setup" and "Running out of the box!".
However…
liu83 updated
1 month ago
-
**Description**
If I loaded 2 model transformer and inference model, memory GPU used about 3Gi.
```
PID USER DEV TYPE GPU GPU MEM CPU HOST MEM Command
2207044 coreai 0 C…