-
### Question
I
It report bug:
-
Hi, thanks for sharing the great work.
But I have some problems when reproducing the results of T5-base on Spider dataset with prefix tuning. I use the following command with 8 gpus:
python -m torch…
-
I'm getting nan values for training and validation loss with recent git versions of mlx and mlx_lm (see commit hashes below).
This is the configuration file I'm using:
```yaml
config.yaml:
mod…
-
Hi, thanks for your great contributions.
I am curious about your ``pretrain-finetune'' pipeline.
According to the paper and your code, it seems that the pipeline is:
1) you first carry out pre-…
-
**Submitting author:** @juanmc2005 (Juan Manuel Coria)
**Repository:** https://github.com/juanmc2005/StreamingSpeakerDiarization
**Branch with paper.md** (empty if default branch): joss
**Version:** v…
-
ubuntu@ip-172-31-72-127:~$ python3 /home/ubuntu/ChatGLM-Efficient-Tuning/src/cli_demo.py \
> --checkpoint_dir /home/ubuntu/p-t-chatglm2v3_\
> --model_name_or_path /home/ubuntu/chatglm2_v3\
…
-
### Model Name
DRKG_COVID19
### Model Description
Drug-Repurposing for COVID-19
### Slug
COVID-19-Drug-Repurposing
### Tag
COVID-19, Drug Repurposing Knowledge Graph
### Publication
https://a…
-
Just a little suggestion, adding some Parameter-Efficient Fine-Tuning (PEFT) methods related to LLM would be nice. One such reference is https://github.com/huggingface/peft
Thanks for this excelle…
-
When i run the finetuning code with --quant for efficient training, I got an error " Integer parameters are unsupported" in torch/distributed/fsdp/flat_param.py", line 435, in _init_flat_param.
Is …
-
```
This is to log the UPC non-blocking memory copy library extensions.
For more information, please see
https://sites.google.com/a/lbl.gov/upc-proposals/extending-the-upc-memory-copy-l
ibrary-functi…