zliucr / mixed-language-training

Attention-Informed Mixed-Language Training for Zero-shot Cross-lingual Task-oriented Dialogue Systems (AAAI-2020)
MIT License
31 stars 6 forks source link

Asking for the source code of Transformer/Multi. BERT/XLM #5

Open Jiahuan-Pei opened 2 years ago

Jiahuan-Pei commented 2 years ago

Hi Zihan,

Thanks for contributing your source code of the paper https://arxiv.org/pdf/1911.09273.pdf.

In this source code, I have not find the Transformer/Multi. BERT/XLM models, as they are the state-of-the-art models as reported in the paper.

Could you please share the models or let us know how to reproduce the reported model.

Regards, Jiahuan

Jiahuan-Pei commented 2 years ago

Could you please also share the settings for reproducing the reported results?

This is what we get using your settings in the README file.

Goal ACC: 0.4840. Joint ACC: 0.1306. Request ACC: 0.8092. Avg ACC: 0.4699

zliucr commented 2 years ago

Hi,

In our experiments, using multilingual word embeddings can achieve comparable or sometimes better results than using Multi. BERT/XLM. If you want to use Multi.BERT/XLM, you can simply replace the word embeddings with the Multi. BERT/XLM embeddings.

As for reproducing the results, we have provided the scripts in the README (in "How to run" section).

Thank you!

Jiahuan-Pei commented 2 years ago

Thanks for your kind reply!

As for reproducing the results, we have provided the scripts in the README (in "How to run" section). ==> Yes, we use the the scripts in the README (in "How to run" section), DST Task, Zero-shot Adaptation for German. And the reproduced results are: Goal ACC: 0.4840. Joint ACC: 0.1306. Request ACC: 0.8092. Avg ACC: 0.4699 The results reported in the paper are:

image

As for model, I do not think it's only the problem of embeddings. (1) In the Figure 3 in the paper (https://arxiv.org/pdf/1911.09273.pdf), it shows you used Transformer encoder. So I assume it's not only need to change Multi. BERT/XLM embeddings. (2) In the Table 1, for German sub table, in the paper, it shows you used Model namely, MUSE XLM (MLM)∗ + Transformer XLM (MLM+TLM)∗+ Transformer Multi. BERT∗+ Transformer

image

Could you please provide more details about the Model and settings for reproducing the models?

zliucr commented 2 years ago

Hi,

The script you run should be able to reproduce the results of MUSE model. The hyper-parameter settings are in config.py file. Can you check whether you use the correct embeddings we have provided in the data folder? Thanks.

As for the code for Multi. BERT/XLM + Transformer, we are sorry that we didn't provide the code, since it is a bit messy in our codebase. If you need, we can try to wrap the corresponding code and upload it in next following days.

Thank you!

Jiahuan-Pei commented 2 years ago

Can you check whether you use the correct embeddings we have provided in the data folder? Thanks. ==> Yes, to make sure the reproducibility, we have NOT changed anything and only use all the codes/data in this repository.

Jiahuan-Pei commented 2 years ago

If you need, we can try to wrap the corresponding code and upload it in next following days. ==>Yes, of course :) It would be much more than appreciated! Thank you!

bumbumya commented 2 years ago

@zliucr Thanks for your great work. Can you share the code for Multi. BERT/XLM + Transformer?