You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I used L=3, K=6 WMT2015 de-en Transformer-Base to reproduce the results of the paper, but the bleu value was nearly 1.64 lower than the original paper.
As shown in the figure above, the result in the paper is bleu 29.29, but the reproduced result is bleu 27.65. The training and inference steps are the same as those on github, so I think there may be something wrong with my pre-processing steps.
I used L=3, K=6 WMT2015 de-en Transformer-Base to reproduce the results of the paper, but the bleu value was nearly 1.64 lower than the original paper.
As shown in the figure above, the result in the paper is bleu 29.29, but the reproduced result is bleu 27.65.
The training and inference steps are the same as those on github, so I think there may be something wrong with my pre-processing steps.
Refer to prepare-wmt14en2de.sh (https://github.com/ictnlp/HMT/blob/main/examples/translation/prepare-wmt14en2de.sh), change it to wmt15, and delete lines 114-118; As the paper, I use newstest2013 (3000 pairs) as the validation set and newstest2015 (2169 pairs) as the test set.
and BPE_TOKENS =32000;The rest of the steps remain the same.
python $BPEROOT/learn_bpe.py -s $BPE_TOKENS < $TRAIN > $BPE_CODE
Finally, perform length filtering:
perl $CLEAN -ratio 1.5 $tmp/bpe.train $src $tgt $prep/train 1 250
perl $CLEAN -ratio 1.5 $tmp/bpe.valid $src $tgt $prep/valid 1 250
Before pre-processing:
After pre-processing:
Can you help analyze the problem, or provide a pre-processing script?
The text was updated successfully, but these errors were encountered: