Note 1: Transformer

2020-07-08  本文已影响0人  qin7zhen

Attention Is All You Need [1]

[1]

1. Encoder-Decoder

Overview of Transformer

2. Attention

3. Position-wise Feed-forward Networks

FFN(x)=max(0, xw_i+b_i)w_2+b_2

4. Positional Encoding


Reference

[1] Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., ... & Polosukhin, I. (2017). Attention is all you need. In Advances in neural information processing systems (pp. 5998-6008).
[2] 口仆. Transformer 原理解析 https://zhuanlan.zhihu.com/p/135873679

上一篇 下一篇

猜你喜欢

热点阅读