mirror of
				https://github.com/labmlai/annotated_deep_learning_paper_implementations.git
				synced 2025-10-31 18:58:43 +08:00 
			
		
		
		
	gpt links
This commit is contained in:
		| @ -21,6 +21,7 @@ contains implementations for | |||||||
| and | and | ||||||
| [relative multi-headed attention](https://lab-ml.com/labml_nn/transformers/relative_mha.html). | [relative multi-headed attention](https://lab-ml.com/labml_nn/transformers/relative_mha.html). | ||||||
|  |  | ||||||
|  | * [GPT Architecture](https://lab-ml.com/labml_nn/transformers/gpt) | ||||||
| * [kNN-LM: Generalization through Memorization](https://lab-ml.com/labml_nn/transformers/knn) | * [kNN-LM: Generalization through Memorization](https://lab-ml.com/labml_nn/transformers/knn) | ||||||
| * [Feedback Transformer](https://lab-ml.com/labml_nn/transformers/feedback) | * [Feedback Transformer](https://lab-ml.com/labml_nn/transformers/feedback) | ||||||
|  |  | ||||||
|  | |||||||
| @ -18,6 +18,10 @@ and derivatives and enhancements of it. | |||||||
| * [Transformer Encoder and Decoder Models](models.html) | * [Transformer Encoder and Decoder Models](models.html) | ||||||
| * [Fixed positional encoding](positional_encoding.html) | * [Fixed positional encoding](positional_encoding.html) | ||||||
|  |  | ||||||
|  | ## [GPT Architecture](gpt) | ||||||
|  |  | ||||||
|  | This is an implementation of GPT-2 architecture. | ||||||
|  |  | ||||||
| ## [kNN-LM](knn) | ## [kNN-LM](knn) | ||||||
|  |  | ||||||
| This is an implementation of the paper | This is an implementation of the paper | ||||||
|  | |||||||
| @ -27,6 +27,7 @@ contains implementations for | |||||||
| and | and | ||||||
| [relative multi-headed attention](https://lab-ml.com/labml_nn/transformers/relative_mha.html). | [relative multi-headed attention](https://lab-ml.com/labml_nn/transformers/relative_mha.html). | ||||||
|  |  | ||||||
|  | * [GPT Architecture](https://lab-ml.com/labml_nn/transformers/gpt) | ||||||
| * [kNN-LM: Generalization through Memorization](https://lab-ml.com/labml_nn/transformers/knn) | * [kNN-LM: Generalization through Memorization](https://lab-ml.com/labml_nn/transformers/knn) | ||||||
| * [Feedback Transformer](https://lab-ml.com/labml_nn/transformers/feedback) | * [Feedback Transformer](https://lab-ml.com/labml_nn/transformers/feedback) | ||||||
|  |  | ||||||
|  | |||||||
		Reference in New Issue
	
	Block a user
	 Varuna Jayasiri
					Varuna Jayasiri