Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

question about details of parameters #5

Open
1245244103 opened this issue Jan 28, 2024 · 3 comments
Open

question about details of parameters #5

1245244103 opened this issue Jan 28, 2024 · 3 comments

Comments

@1245244103
Copy link

hello, i am trying to reproduce the result in the paper.I run the scripts/run_epr.sh successfully. i get the em score of about 71 in mrpc which is 75.98 in the paper.Are the settings in run_epr.sh different from the paper? Can you provide the setting parameters in the paper?
Thanks a lot!

@jiacheng-ye
Copy link
Contributor

Hi, it's weird as the settings in run_epr.sh is the same as that in the paper. Could you check whether you can obtain similar results to the paper for other methods such as Topk-BERT as I'm not sure if it's due to the randomness of running on different machines.

@hanxinyan20
Copy link

I found the same question, too. I run the script/run_bm25.sh, and got the acc on mrpc validation set is 0.576 which is far lower than in the paper. I just change the num_ice to 27, and leave other parameters unchanged. I also tried to evaluate sst5 ( set num_ice to 27), the acc is 0.296. Can you give me some advice so that I can get the same acc as you claimed in your paper?

@1245244103
Copy link
Author

I found the same question, too. I run the script/run_bm25.sh, and got the acc on mrpc validation set is 0.576 which is far lower than in the paper. I just change the num_ice to 27, and leave other parameters unchanged. I also tried to evaluate sst5 ( set num_ice to 27), the acc is 0.296. Can you give me some advice so that I can get the same acc as you claimed in your paper?

While replicating the process, I noticed issues with the training code for the encoder. I've rewritten the training code without using a trainer, following the style of Hugging Face, and without the use of accelerate. Additionally, there were some discrepancies between the parameters used in the code and those described in the paper. For instance, the paper mentions using the three samples with the highest and lowest scores as positive and negative examples, respectively, whereas the code only samples one. I have made adjustments to align with the paper. After these modifications, the results on some datasets are close to those reported in the paper. You might want to give it a try.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants