You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm reading your paper and find that the BLEU scores you compared with state-of-the-art models in Table 2 are based on BLEU1. But it seems that they calculate BLEU scores based on BLEU4, which is suggested in the original BLEU paper "Bleu: a method for automatic evaluation of machine translation".
Another paper published in ACL 2019 seems to use BLEU4 too. "Generating Sentences from Disentangled Syntactic and Semantic Spaces". As shown in their published codes.
So I'm wondering whether the BLEU scores are the same.
The text was updated successfully, but these errors were encountered:
Hi,
I'm reading your paper and find that the BLEU scores you compared with state-of-the-art models in Table 2 are based on BLEU1. But it seems that they calculate BLEU scores based on BLEU4, which is suggested in the original BLEU paper "Bleu: a method for automatic evaluation of machine translation".
Another paper published in ACL 2019 seems to use BLEU4 too. "Generating Sentences from Disentangled Syntactic and Semantic Spaces". As shown in their published codes.
So I'm wondering whether the BLEU scores are the same.
The text was updated successfully, but these errors were encountered: