Sequence-to-Sequence (seq2seq) Neural Network with Attention

From GM-RKB
Jump to navigation Jump to search

A Sequence-to-Sequence (seq2seq) Neural Network with Attention is a neural seq2seq network that includes an attention mechanism.



References

2017a

Figure 3: Pointer-generator model. For each decoder timestep a generation probability $p_{gen} \in [0,1]$ is calculated, which weights the probability of generating words from the vocabulary, versus copying words from the source text. The vocabulary distribution and the attention distribution are weighted and summed to obtain the final distribution, from which we make our prediction. Note that out-of-vocabulary article words such as 2-0 are included in the final distribution. Best viewed in color.

2017b

2015a

2015b