Attention Mechanism

What is the name of the machine learning technique that allows a neural network to focus on specific parts of an input sequence?check

Attention mechanism

How does an attention model differ from a traditional model?

Attention models pass a lot more information to the decoder.

What is the name of the machine learning architecture that can be used to translate text from one language to another?check

Encoder-decoder

What is the advantage of using the attention mechanism over a traditional recurrent neural network (RNN) encoder-decoder?

The attention mechanism lets the decoder focus on specific parts of the input sequence, which can improve the accuracy of the translation.

What is the advantage of using the attention mechanism over a traditional sequence-to-sequence model?

The attention mechanism lets the model focus on specific parts of the input sequence.

What is the purpose of the attention weights?check

To assign weights to different parts of the input sequence, with the most important parts receiving the highest weights.

What are the two main steps of the attention mechanism?

Calculating the attention weights and generating the context vector

Unknown's avatar

Author: Ajay Ohri

http://about.me/ajayohri

Leave a comment