top of page

What it takes to learn transformer models for natural langauge processing!-part II

joydeepml2020

In my last post, we have discussed about simple encoder decoder based seq2seq models.

In this article, we will try to understand the shortcommings of simple encoder-decoder based model and what solution is bought by attention models.



 
 
 

Commentaires


bottom of page