Open heya5 opened 2 years ago
I have just solved that problem on triviaQA. In the paper's setting, the reranker's training epoch is 1 and the warmup is 10%. So the second to last line should be --num_train_epochs 1 --warmup_fraction 0.1 \ I use that command to reproduce on triviaQA and get a similar result(even higher than the result in the paper). wish this can help you!
@Fu-Dayuan When you reproduce the Trivia QA dataset, do you remember the results of the RAG model in the second stage (in this stage, the generation model only uses the retrieved passages from DPR, very similar to KGI). The training command is:
python generation/kgi_train.py \
--kilt_data ${dataset}_training \
--output models/RAG/${dataset}_dpr_rag \
--corpus_endpoint kilt_passages_${dataset} \
--model_name facebook/rag-token-nq \
--model_path models/RAG/${dataset}_dpr_rag_init \
--warmup_fraction 0.05 --num_train_epochs 2 \
--learning_rate 3e-5 --full_train_batch_size 128 --gradient_accumulation_steps 64
I got the following performance | R-Prec | Recall@5 | Accuracy | F1 | KILT-AC | KILT-F1 |
---|---|---|---|---|---|---|
578 | .621 | .375 | .573 | .261 | .385 |
The retrieval metrics are close to the KGI performance, but generation metrics are far worse than the KGI performance.
kgi_train.py
.But after I run
reranker_train.py
andrerank_apply.py
, I get a results which seems worse than the results in the table 2 of Re2G paper.I think re2g is a solid work. Could you please give me some advices to reproduce the result?
I run the following command to train the reranker