Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about the BLEU scores compared with other state-of-the-art models #6

Open
chenwq95 opened this issue Oct 25, 2019 · 0 comments

Comments

@chenwq95
Copy link

Hi,

I'm reading your paper and find that the BLEU scores you compared with state-of-the-art models in Table 2 are based on BLEU1. But it seems that they calculate BLEU scores based on BLEU4, which is suggested in the original BLEU paper "Bleu: a method for automatic evaluation of machine translation".

Another paper published in ACL 2019 seems to use BLEU4 too. "Generating Sentences from Disentangled Syntactic and Semantic Spaces". As shown in their published codes.

So I'm wondering whether the BLEU scores are the same.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant