mirror of
https://github.com/labmlai/annotated_deep_learning_paper_implementations.git
synced 2025-10-29 09:38:56 +08:00
experiment links transformer
This commit is contained in:
@ -11,6 +11,7 @@
|
||||
"source": [
|
||||
"[](https://github.com/labmlai/annotated_deep_learning_paper_implementations)\n",
|
||||
"[](https://colab.research.google.com/github/labmlai/annotated_deep_learning_paper_implementations/blob/master/labml_nn/transformers/basic/autoregressive_experiment.ipynb)\n",
|
||||
"[](comet.ml/labml/transformer/ea8c108c2d94434ca3c2bc2b21015082)\n",
|
||||
"\n",
|
||||
"## Transformer Experiment\n",
|
||||
"\n",
|
||||
|
||||
@ -7,10 +7,11 @@ summary: >
|
||||
|
||||
# Transformer Auto-Regression Experiment
|
||||
|
||||
[](https://colab.research.google.com/github/labmlai/annotated_deep_learning_paper_implementations/blob/master/labml_nn/transformers/basic/autoregressive_experiment.ipynb)
|
||||
[](comet.ml/labml/transformer/ea8c108c2d94434ca3c2bc2b21015082)
|
||||
|
||||
This trains a simple transformer introduced in [Attention Is All You Need](https://papers.labml.ai/paper/1706.03762)
|
||||
on an NLP auto-regression task (with Tiny Shakespeare dataset).
|
||||
|
||||
[](https://colab.research.google.com/github/labmlai/annotated_deep_learning_paper_implementations/blob/master/labml_nn/transformers/basic/autoregressive_experiment.ipynb)
|
||||
"""
|
||||
|
||||
import torch
|
||||
|
||||
@ -8,6 +8,9 @@ summary: >
|
||||
|
||||
# Multi-Headed Attention (MHA)
|
||||
|
||||
[](https://colab.research.google.com/github/labmlai/annotated_deep_learning_paper_implementations/blob/master/labml_nn/transformers/basic/autoregressive_experiment.ipynb)
|
||||
[](comet.ml/labml/transformer/ea8c108c2d94434ca3c2bc2b21015082)
|
||||
|
||||
This is a tutorial/implementation of multi-headed attention
|
||||
from paper [Attention Is All You Need](https://papers.labml.ai/paper/1706.03762)
|
||||
in [PyTorch](https://pytorch.org/).
|
||||
@ -17,8 +20,6 @@ Here is the [training code](basic/autoregressive_experiment.html) that uses a ba
|
||||
with MHA for NLP auto-regression.
|
||||
|
||||
[Here is an experiment implementation](basic/autoregressive_experiment.html) that trains a simple transformer.
|
||||
|
||||
[](https://colab.research.google.com/github/labmlai/annotated_deep_learning_paper_implementations/blob/master/labml_nn/transformers/basic/autoregressive_experiment.ipynb)
|
||||
"""
|
||||
|
||||
import math
|
||||
|
||||
@ -7,6 +7,9 @@ summary: >
|
||||
---
|
||||
|
||||
# Transformer Encoder and Decoder Models
|
||||
|
||||
[](https://colab.research.google.com/github/labmlai/annotated_deep_learning_paper_implementations/blob/master/labml_nn/transformers/basic/autoregressive_experiment.ipynb)
|
||||
[](comet.ml/labml/transformer/ea8c108c2d94434ca3c2bc2b21015082)
|
||||
"""
|
||||
import math
|
||||
|
||||
|
||||
Reference in New Issue
Block a user