반응형
deeplearning
-
Transformer EncoderML engineer/NLP 2022. 12. 31. 01:00
🕓 6 mins read By the time anyone has reached this post, chances are you already know pretty much all about Transformer models. It's no wonder since the paper Attention is all you need is already an old paper from 2017, which means it's pretty ancient in this field. Well I take should take that back. It's not that ancient.. considering how not much has changed in the seq2seq model paradigm. (GPT3..