In this lecture, we will look at how the idea of attention can be used in NLP contexts. We will first see how attention can be used with encoder-decoder networks. Then we will look at variants of the idea that generalizes it beyond sequential inputs or outputs.
Deep Learning for NLP
CS 6956, Spring 2019