Open Yazooliu opened 1 year ago
You can set export CUDA_VISIBLE_DEVICES=1
to use GPU 1. The detail can refer to https://stackoverflow.com/questions/39649102/how-do-i-select-which-gpu-to-run-a-job-on .
You can set
export CUDA_VISIBLE_DEVICES=1
to use GPU 1. The detail can refer to https://stackoverflow.com/questions/39649102/how-do-i-select-which-gpu-to-run-a-job-on .
other issue occur:
delete --fp16
delete
--fp16
yes. it works. but maybe some issue in code as following????
this is code:
this is logger output info and error detail, (1) logger output info
09/26/2023 16:47:07 - WARNING - main - Process rank: 0, device: cpu, n_gpu: 0, distributed training: True, 16-bits training: False
09/26/2023 16:47:07 - INFO - main - Training/evaluation parameters RetrieverTrainingArguments(
_n_gpu=0,
adafactor=False,
adam_beta1=0.9,
adam_beta2=0.999,
adam_epsilon=1e-08,
auto_find_batch_size=False,
bf16=False,
bf16_full_eval=False,
data_seed=None,
dataloader_drop_last=True,
dataloader_num_workers=0,
dataloader_pin_memory=True,
ddp_backend=None,
ddp_broadcast_buffers=None,
ddp_bucket_cap_mb=None,
ddp_find_unused_parameters=None,
ddp_timeout=1800,
debug=[],
deepspeed=None,
disable_tqdm=False,
dispatch_batches=None,
do_eval=False,
do_predict=False,
do_train=False,
eval_accumulation_steps=None,
eval_delay=0,
eval_steps=None,
evaluation_strategy=no,
fix_position_embedding=False,
fp16=False,
fp16_backend=auto,
fp16_full_eval=False,
fp16_opt_level=O1,
fsdp=[],
fsdp_config={'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False},
fsdp_min_num_params=0,
fsdp_transformer_layer_cls_to_wrap=None,
full_determinism=False,
gradient_accumulation_steps=1,
gradient_checkpointing=False,
greater_is_better=None,
group_by_length=False,
half_precision_backend=auto,
hub_always_push=False,
hub_model_id=None,
hub_private_repo=False,
hub_strategy=every_save,
hub_token=
error info:
It seems that 'RetrieverTrainingArguments' object has no attribute 'sentence_pooling_method'
You can use the latest code. git clone the repo or pip install -U FlagEmbedding
You can use the latest code. git clone the repo or pip install -U FlagEmbedding
seems your code support 2 diff generated embedding method. which will be better?
For bge, cls
maybe better because we use cls in training. However, you can select different settings based on the performance of your data.
For bge,
cls
maybe better because we use cls in training. However, you can select different settings based on the performance of your data.
If I hava any good change or idea, could I support to PR? Any limitaiton from your side?
You can use the latest code. git clone the repo or pip install -U FlagEmbedding your this pip install method can not support this:
For bge,
cls
maybe better because we use cls in training. However, you can select different settings based on the performance of your data.If I hava any good change or idea, could I support to PR? Any limitaiton from your side?
Sure. PR is very welcome, and there is no limitation.
You can use the latest code. git clone the repo or pip install -U FlagEmbedding your this pip install method can not support this:
Sorry, I forgot to add the sentence_transformers to setup.py. You can install sentence_transformers manually.
After prepare the training env , I try to finetune the model as following the step2 and step3 in https://github.com/FlagOpen/FlagEmbedding/tree/master/examples/finetune#hard-negatives
step2 is done and can generated the xxxHN.jsonl by following step3, it will raise the CUDA of memorry. My GPU1 memory is full but GPU2 can ok. How I can to edit the code or change to us one GPU to finetune.
My GPU info as following:
and trianing command as:
or do you hava any good suggetion to us less GPU resource or set to GPU No.2 to run it successfully? I only have 70 query with positive and negative example to finetune as the data is too less.
Thanks for help.