Open WQi777 opened 2 years ago
Hi WQi777, Thanks for your interest. We just added a line to reproduce results of DeBERTa v3 to https://github.com/microsoft/KEAR/blob/main/bash/task_train.sh. Hope that helps!
Thanks a lot for your reply!
but when i run the code ,got an error:
batch size: 4, total_batch_size: 20
[1528]: world_size = 2, rank = 1, backend=nccl
batch size: 4, total_batch_size: 20
restarting from checkpoint.
used_name: last2
restarting from checkpoint.
used_name: last2
loading result from dir test/last2
args.fp16 is 0
loading result from dir test/last2
args.fp16 is 0
load_vocab microsoft/deberta-v3-large
load_vocab microsoft/deberta-v3-large
Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
load_data data/csqa_ret_3datasets/train_data.json
Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
load_data data/csqa_ret_3datasets/train_data.json
data: 9741, world_size: 2
load_data data/csqa_ret_3datasets/dev_data.json
data: 1222, world_size: 2
get dir test/
make dataloader ...
data: 9741, world_size: 2
load_data data/csqa_ret_3datasets/dev_data.json
data: 1222, world_size: 2
get dir test/
make dataloader ...
max len: 968
95 percent len: 490
train_data 9741
total length: 1218
max len: 968
95 percent len: 490
train_data 9741
total length: 1218
max len: 851
95 percent len: 514
devlp_data 1222
init_model test/last2
set config, model_type= debertav2
deepspeed: True
resume_training: True
config_path:test/last2
model_type= debertav2
Traceback (most recent call last):
File "task.py", line 409, in
Looking forward to your reply.
Hi WQi777, We have a typo in our code - can you try again?
Hello, excuse me.Would you tell me how can I reproduce the results in your paper? when I train the model according to the method in the ’readme‘, the accuracy rate I get continues to drop with each round of training. Can you tell me what is the reason?Looking forward to your reply~