Coursera, Deep Learning 5, Sequence Models, week4, Transformer Network

 self-attention

 multi-head attention

转载请注明出处 http://www.cnblogs.com/mashuai-191/
原文地址:https://www.cnblogs.com/mashuai-191/p/15532600.html