Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Asking for the source code of Transformer/Multi. BERT/XLM #5

Open
Jiahuan-Pei opened this issue Mar 22, 2022 · 7 comments
Open

Asking for the source code of Transformer/Multi. BERT/XLM #5

Jiahuan-Pei opened this issue Mar 22, 2022 · 7 comments

Comments

@Jiahuan-Pei
Copy link

Hi Zihan,

Thanks for contributing your source code of the paper https://arxiv.org/pdf/1911.09273.pdf.

In this source code, I have not find the Transformer/Multi. BERT/XLM models, as they are the state-of-the-art models as reported in the paper.

Could you please share the models or let us know how to reproduce the reported model.

Regards,
Jiahuan

@Jiahuan-Pei
Copy link
Author

Could you please also share the settings for reproducing the reported results?

This is what we get using your settings in the README file.

Goal ACC: 0.4840. Joint ACC: 0.1306. Request ACC: 0.8092. Avg ACC: 0.4699

@zliucr
Copy link
Owner

zliucr commented Mar 22, 2022

Hi,

In our experiments, using multilingual word embeddings can achieve comparable or sometimes better results than using Multi. BERT/XLM. If you want to use Multi.BERT/XLM, you can simply replace the word embeddings with the Multi. BERT/XLM embeddings.

As for reproducing the results, we have provided the scripts in the README (in "How to run" section).

Thank you!

@Jiahuan-Pei
Copy link
Author

Thanks for your kind reply!

As for reproducing the results, we have provided the scripts in the README (in "How to run" section).
==> Yes, we use the the scripts in the README (in "How to run" section), DST Task, Zero-shot Adaptation for German. And the reproduced results are:
Goal ACC: 0.4840. Joint ACC: 0.1306. Request ACC: 0.8092. Avg ACC: 0.4699
The results reported in the paper are:
image

As for model, I do not think it's only the problem of embeddings.
(1) In the Figure 3 in the paper (https://arxiv.org/pdf/1911.09273.pdf), it shows you used Transformer encoder. So I assume it's not only need to change Multi. BERT/XLM embeddings.
(2) In the Table 1, for German sub table, in the paper, it shows you used Model namely,
MUSE
XLM (MLM)∗ + Transformer
XLM (MLM+TLM)∗+ Transformer
Multi. BERT∗+ Transformer
image

Could you please provide more details about the Model and settings for reproducing the models?

@zliucr
Copy link
Owner

zliucr commented Mar 22, 2022

Hi,

The script you run should be able to reproduce the results of MUSE model. The hyper-parameter settings are in config.py file. Can you check whether you use the correct embeddings we have provided in the data folder? Thanks.

As for the code for Multi. BERT/XLM + Transformer, we are sorry that we didn't provide the code, since it is a bit messy in our codebase. If you need, we can try to wrap the corresponding code and upload it in next following days.

Thank you!

@Jiahuan-Pei
Copy link
Author

Jiahuan-Pei commented Mar 22, 2022

Can you check whether you use the correct embeddings we have provided in the data folder? Thanks.
==> Yes, to make sure the reproducibility, we have NOT changed anything and only use all the codes/data in this repository.

@Jiahuan-Pei
Copy link
Author

If you need, we can try to wrap the corresponding code and upload it in next following days.
==>Yes, of course :) It would be much more than appreciated! Thank you!

@bumbumya
Copy link

bumbumya commented Jun 8, 2022

@zliucr
Thanks for your great work.
Can you share the code for Multi. BERT/XLM + Transformer?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants