mirror of
				https://github.com/labmlai/annotated_deep_learning_paper_implementations.git
				synced 2025-10-31 02:39:16 +08:00 
			
		
		
		
	links
This commit is contained in:
		| @ -36,6 +36,7 @@ implementations. | ||||
| * [Pay Attention to MLPs (gMLP)](transformers/gmlp/index.html) | ||||
| * [Vision Transformer (ViT)](transformers/vit/index.html) | ||||
| * [Primer EZ](transformers/primer_ez/index.html) | ||||
| * [Hourglass](transformers/hourglass/index.html) | ||||
|  | ||||
| #### ✨ [Recurrent Highway Networks](recurrent_highway_networks/index.html) | ||||
|  | ||||
|  | ||||
| @ -93,6 +93,11 @@ This is an implementation of the paper | ||||
|  | ||||
| This is an implementation of the paper | ||||
| [Primer: Searching for Efficient Transformers for Language Modeling](https://papers.labml.ai/paper/2109.08668). | ||||
|  | ||||
| ## [Hourglass](hour_glass/index.html) | ||||
|  | ||||
| This is an implementation of the paper | ||||
| [Hierarchical Transformers Are More Efficient Language Models](https://papers.labml.ai/paper/2110.13711) | ||||
| """ | ||||
|  | ||||
| from .configs import TransformerConfigs | ||||
|  | ||||
		Reference in New Issue
	
	Block a user
	 Varuna Jayasiri
					Varuna Jayasiri