In this lecture, we will look at how the idea of attention can be used in NLP contexts. We will first see how attention can be used with encoder-decoder networks. Then we will look at variants of the idea that generalizes it beyond sequential inputs or outputs.
NLP with Neural Networks
CS 6957, Fall 2023