gpt links

This commit is contained in:
Varuna Jayasiri
2021-01-14 10:20:21 +05:30
parent 2cf6e7a079
commit a94ed927aa
3 changed files with 6 additions and 0 deletions

View File

@ -21,6 +21,7 @@ contains implementations for
and and
[relative multi-headed attention](https://lab-ml.com/labml_nn/transformers/relative_mha.html). [relative multi-headed attention](https://lab-ml.com/labml_nn/transformers/relative_mha.html).
* [GPT Architecture](https://lab-ml.com/labml_nn/transformers/gpt)
* [kNN-LM: Generalization through Memorization](https://lab-ml.com/labml_nn/transformers/knn) * [kNN-LM: Generalization through Memorization](https://lab-ml.com/labml_nn/transformers/knn)
* [Feedback Transformer](https://lab-ml.com/labml_nn/transformers/feedback) * [Feedback Transformer](https://lab-ml.com/labml_nn/transformers/feedback)

View File

@ -18,6 +18,10 @@ and derivatives and enhancements of it.
* [Transformer Encoder and Decoder Models](models.html) * [Transformer Encoder and Decoder Models](models.html)
* [Fixed positional encoding](positional_encoding.html) * [Fixed positional encoding](positional_encoding.html)
## [GPT Architecture](gpt)
This is an implementation of GPT-2 architecture.
## [kNN-LM](knn) ## [kNN-LM](knn)
This is an implementation of the paper This is an implementation of the paper

View File

@ -27,6 +27,7 @@ contains implementations for
and and
[relative multi-headed attention](https://lab-ml.com/labml_nn/transformers/relative_mha.html). [relative multi-headed attention](https://lab-ml.com/labml_nn/transformers/relative_mha.html).
* [GPT Architecture](https://lab-ml.com/labml_nn/transformers/gpt)
* [kNN-LM: Generalization through Memorization](https://lab-ml.com/labml_nn/transformers/knn) * [kNN-LM: Generalization through Memorization](https://lab-ml.com/labml_nn/transformers/knn)
* [Feedback Transformer](https://lab-ml.com/labml_nn/transformers/feedback) * [Feedback Transformer](https://lab-ml.com/labml_nn/transformers/feedback)