top of page
joydeepml2020

What it takes to learn transformer models for natural langauge processing!-part II

In my last post, we have discussed about simple encoder decoder based seq2seq models.

In this article, we will try to understand the shortcommings of simple encoder-decoder based model and what solution is bought by attention models.



10 views0 comments

Comments


bottom of page