Exploiting Cross-Sentence Context for Neural Machine Translation
Longyue Wang♥ Zhaopeng Tu♠ Andy Way♥ Qun Liu♥
♥ ADAPT Centre, Dublin City University ♠ Tencent AI Lab
Exploiting Cross-Sentence Context for Neural Machine Translation - - PowerPoint PPT Presentation
Exploiting Cross-Sentence Context for Neural Machine Translation Longyue Wang Zhaopeng Tu Andy Way Qun Liu ADAPT Centre, Dublin City University Tencent AI Lab Motivation The
♥ ADAPT Centre, Dublin City University ♠ Tencent AI Lab
⨁
这是 ⼀丁个 ⽣甠态 ⽹罒络 。 <eos> 0.1 0.2 0.7 0.1 0.0 0.0
c st s1 s0 this ct a neutral sT cT . <eos>
(Choi et al., 2016)
Word x0 Axis Nearest Neighbours notebook 1 diary notebooks (notebook) sketchbook jottings 2 palmtop notebooks (notebook) ipaq laptop power 1 powers authority (power) powerbase sovereignity 2 powers electrohydraulic microwatts hydel (power)
Past 那么 在 这个 问题 上 , 伊朗 的 … well, on this issue , iran has a relatively … 在 任内 解决 伊朗 核 问题 , 不泌管是 ⽤甩 和平 … to resolve the iranian nuclear issue in his term , … Current 那 刚刚 提到 这个 … 谈判 的 问题 。 that just mentioned the issue of the talks …
这⾥里離 … ⾃臫然
保护区
。 <eos> … 我们 …
⼴庀义的
养殖 。 <eos> … … … …
Word-level RNN Sentence-level RNN Cross-Sentence Context
Cross-Sentence Context
这是 ⼀丁个 ⽣甠态 ⽹罒络 。 <eos>
Cross-Sentence Context
这是 ⼀丁个 ⽣甠态 ⽹罒络 。 <eos>
st s1 s0 sT
Cross-Sentence Context
这是 ⼀丁个 ⽣甠态 ⽹罒络 。 <eos>
st s1 s0 sT
32.0 31.9 31.55 30.57
Baseline +Init_Enc +Init_Dec +Init_Both
Cross-Sentence Context
⨁
这是 ⼀丁个 ⽣甠态 ⽹罒络 。 <eos> 0.1 0.2 0.7 0.1 0.0 0.0
st s1 s0 this ct a neutral sT . <eos>
Intra-Sentence Context
ct st-1 yt-1 D 𝜏
✕
𝑨t st-1 yt-1 D ct st-1 yt-1
(a) standard decoder (b) decoder with auxiliary context (c) decoder with gating auxiliary context
ct st
act.
st
act.
st
act.
32.24 31.3 30.57
Baseline +Aux. Ctx. +Gating Aux. Ctx.
Cross-Sentence Context
⨁
这是 ⼀丁个 ⽣甠态 ⽹罒络 。 <eos> 0.1 0.2 0.7 0.1 0.0 0.0
st s1 s0 this ct a neutral sT . <eos>
Intra-Sentence Context
32.67 32.24 32.00 30.57
Baseline +Init_Both +Gating Aux. Ctx. +Both
Hist. Ÿ Ié @ –M J… *ò Ï v' l˚ j¡ ⌫ ? Input ˝& O6 å ⌥Q Pò ? Ref. Can it inhibit and deter corrupt offi- cials? NMT Can we contain and deter the enemy? Our Can it contain and deter the corrupt
representation for neural machine translation. arXiv 2016.
Simonsen, and Jian- Yun Nie. A hierarchical recurrent encoder- decoder for generative context-aware query suggestion. CIKM 2015.
network models. AAAI 2016.
International Conference on Machine Learning, Deep Learning Workshop.