|
|
|
@ -4,6 +4,12 @@
|
|
|
|
|
|
|
|
|
|
有疑惑的地方欢迎issue或邮件我,😀Enjoin!
|
|
|
|
|
|
|
|
|
|
[Attention Is All You Need 论文地址](https://arxiv.org/pdf/1706.03762)
|
|
|
|
|
|
|
|
|
|
[GPT-2 官方代码](https://www.tensorflow.org/tutorials/text/transformer?hl=zh-cn)
|
|
|
|
|
|
|
|
|
|
[动态图解 nanoGPT](https://bbycroft.net/llm)
|
|
|
|
|
|
|
|
|
|
### 学习章节:
|
|
|
|
|
|
|
|
|
|
<ul>
|
|
|
|
@ -17,3 +23,4 @@
|
|
|
|
|
<li><a href='https://github.com/ben1234560/AiLearning-Theory-Applying/blob/master/%E4%BA%BA%E4%BA%BA%E9%83%BD%E8%83%BD%E7%9C%8B%E6%87%82%E7%9A%84Transformer/%E7%AC%AC%E5%85%AD%E7%AB%A0%E2%80%94%E2%80%94%E6%95%B0%E5%80%BC%E7%BC%A9%E6%94%BE.md'>第六章——数值缩放</a>
|
|
|
|
|
<li><a href='https://github.com/ben1234560/AiLearning-Theory-Applying/blob/master/%E4%BA%BA%E4%BA%BA%E9%83%BD%E8%83%BD%E7%9C%8B%E6%87%82%E7%9A%84Transformer/%E7%AC%AC%E4%B8%83%E7%AB%A0%E2%80%94%E2%80%94%E5%89%8D%E9%A6%88%E7%A5%9E%E7%BB%8F%E7%BD%91%E7%BB%9C.md'>第七章——前馈神经网络</a>
|
|
|
|
|
<li><a href='https://github.com/ben1234560/AiLearning-Theory-Applying/blob/master/%E4%BA%BA%E4%BA%BA%E9%83%BD%E8%83%BD%E7%9C%8B%E6%87%82%E7%9A%84Transformer/%E7%AC%AC%E5%85%AB%E7%AB%A0%E2%80%94%E2%80%94%E6%9C%80%E5%90%8E%E7%9A%84%E8%BE%93%E5%87%BA.md'>第八章——最后的输出</a>
|
|
|
|
|
|
|
|
|
|