WMT的英德翻译

2019-02-12  本文已影响0人  JackHorse

1. The University of Cambridge’s Machine Translation Systems for WMT18

1. basic Architecture

Combine the three most commonly used architectures: recurrent, convolutional, and self-attention-based models like the Transformer

2. system combination

If we want to combine q models M_1,...,M_q, we first divide the models into two groups by selecting a p with 1 \le p \le q.

Then, we refer to the first group M_1,...,M_p as full posterior scores and the second group M_p,...,M_q as MBR-based scores.

Full-posterior models scores compute as follows:


在这里插入图片描述

Combined scores compute as follows:


在这里插入图片描述

3. Data

1. language detection (Nakatani, 2010) on all available monolingual and parallel data
2. additionally filtered on ParaCrawl

2. NTT’s Neural Machine Translation Systems for WMT 2018

1. basic Architecture

Transformer Big

2. Data

  1. use language model (such as KenLM) to evaluate a sentences naturalness
  2. use a word alignment model (such as fast_align) to check whether the sentence pair has the same meaning
  1. translating monolingual sentences with Transformer -> seudo-parallel corpora
  2. Back-translate & evaluate -> selected the high-scoring sentence pair
  1. R2L model re-ranks an n-best hypothesis generated by the Left-to-Right (L2R) model (n=10)

3. Microsoft’s Submission to the WMT2018 News Translation Task:How I Learned to Stop Worrying and Love the Data

1. basic Architecture

Transformer Big + Ensemble-decoding + R2L Reranking

2. Data

上一篇 下一篇

猜你喜欢

热点阅读