mirror of
https://github.com/labmlai/annotated_deep_learning_paper_implementations.git
synced 2025-08-06 15:22:21 +08:00
rope links
This commit is contained in:
@ -22,6 +22,7 @@ implementations.
|
|||||||
* [Transformer building blocks](transformers/models.html)
|
* [Transformer building blocks](transformers/models.html)
|
||||||
* [Transformer XL](transformers/xl/index.html)
|
* [Transformer XL](transformers/xl/index.html)
|
||||||
* [Relative multi-headed attention](transformers/xl/relative_mha.html)
|
* [Relative multi-headed attention](transformers/xl/relative_mha.html)
|
||||||
|
* [Rotary Positional Embeddings](transformers/rope/index.html)
|
||||||
* [Compressive Transformer](transformers/compressive/index.html)
|
* [Compressive Transformer](transformers/compressive/index.html)
|
||||||
* [GPT Architecture](transformers/gpt/index.html)
|
* [GPT Architecture](transformers/gpt/index.html)
|
||||||
* [GLU Variants](transformers/glu_variants/simple.html)
|
* [GLU Variants](transformers/glu_variants/simple.html)
|
||||||
|
@ -22,6 +22,9 @@ and derivatives and enhancements of it.
|
|||||||
This implements Transformer XL model using
|
This implements Transformer XL model using
|
||||||
[relative multi-head attention](xl/relative_mha.html)
|
[relative multi-head attention](xl/relative_mha.html)
|
||||||
|
|
||||||
|
## [Rotary Positional Embeddings](rope/index.html)
|
||||||
|
This implements Rotary Positional Embeddings (RoPE)
|
||||||
|
|
||||||
## [Compressive Transformer](compressive/index.html)
|
## [Compressive Transformer](compressive/index.html)
|
||||||
|
|
||||||
This is an implementation of compressive transformer
|
This is an implementation of compressive transformer
|
||||||
|
@ -24,6 +24,7 @@ implementations almost weekly.
|
|||||||
* [Transformer building blocks](https://nn.labml.ai/transformers/models.html)
|
* [Transformer building blocks](https://nn.labml.ai/transformers/models.html)
|
||||||
* [Transformer XL](https://nn.labml.ai/transformers/xl/index.html)
|
* [Transformer XL](https://nn.labml.ai/transformers/xl/index.html)
|
||||||
* [Relative multi-headed attention](https://nn.labml.ai/transformers/xl/relative_mha.html)
|
* [Relative multi-headed attention](https://nn.labml.ai/transformers/xl/relative_mha.html)
|
||||||
|
* [Rotary Positional Embeddings](https://nn.labml.ai/transformers/rope/index.html)
|
||||||
* [Compressive Transformer](https://nn.labml.ai/transformers/compressive/index.html)
|
* [Compressive Transformer](https://nn.labml.ai/transformers/compressive/index.html)
|
||||||
* [GPT Architecture](https://nn.labml.ai/transformers/gpt/index.html)
|
* [GPT Architecture](https://nn.labml.ai/transformers/gpt/index.html)
|
||||||
* [GLU Variants](https://nn.labml.ai/transformers/glu_variants/simple.html)
|
* [GLU Variants](https://nn.labml.ai/transformers/glu_variants/simple.html)
|
||||||
|
Reference in New Issue
Block a user