Paper notes-neural machine translation by jointly learning to Align and Translate

Source: Internet
Author: User
Tags wrapper

A attention mechanism is proposed for machine translation.

Background: RNN-based machine translation

The basic idea is to encode the language X first encoder, then decode decoder for language Y. Encoder and decoder can be seen as two layers of rnn? Coded hidden layer h and decoding the hidden layer s

RNN Encoder-decoder:

1) The input sentence, expressed as , through the loop layer to get hidden layer , the vector c is represented as a hidden layer of the function, C is the input encode out of the vector.

2) Next is the decoder stage, predicting the next word based on the previously predicted translated words and the input encoder

The innovation of this paper:

The conditional probabilities of the (2) type are rewritten, and the encoder of each yi,context are different and recorded as CI

About CI calculation: CI is represented as a series of hi linear weights, where hi is the hidden layer of the encoder end, defined as Annotation,hi (personal understanding) as the input of the word "I" near the information (the expression of the input terminal I is simply)

Alpha factor:

Alpha or E represents the annotation of the J-input word and the importance of the i-1 hidden state of the decoder end, so that the resulting CI will pay attention for some locations, equivalently as the translation word I to the original input some position pay Attetnion

Using BIRNN:

This paper uses bidirectional rnn to catch the forward and backward hi stitching together, so that the annotation can represent the information around the input word I.

Network structure:

Paper notes-neural machine translation by jointly learning to Align and Translate

Contact Us

The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion; products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the content of the page makes you feel confusing, please write us an email, we will handle the problem within 5 days after receiving your email.

If you find any instances of plagiarism from the community, please send an email to: info-contact@alibabacloud.com and provide relevant evidence. A staff member will contact you within 5 working days.

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.