top of page
  • joydeepml2020

What it takes to learn transformer models for natural langauge processing!-part II

In my last post, we have discussed about simple encoder decoder based seq2seq models.

In this article, we will try to understand the shortcommings of simple encoder-decoder based model and what solution is bought by attention models.

Sequence2Sequence Models with Attention
Download • 181KB

10 views0 comments


bottom of page