728x90
반응형
'DL > NLP' 카테고리의 다른 글
Transformer 4 - GPT / BERT / 그 외 파생모델들 (0) | 2023.03.14 |
---|---|
Transformer 3 - Residential Add / Normalization (0) | 2023.03.14 |
Transformer 2 - Multi Head Attention / Scaled Dot-Product Attention / Masking / Position-wise Feed Forward Network / Query, Key, Value (0) | 2023.03.14 |
Transformer 1 - Positional Encoding (0) | 2023.03.09 |
모델 발전과정 3 - GNMT (Google's Seq2seq 8 layers w. Residual) (0) | 2023.02.28 |
댓글