mirror of
				https://github.com/labmlai/annotated_deep_learning_paper_implementations.git
				synced 2025-10-31 10:48:49 +08:00 
			
		
		
		
	gpt links
This commit is contained in:
		| @ -21,6 +21,7 @@ contains implementations for | ||||
| and | ||||
| [relative multi-headed attention](https://lab-ml.com/labml_nn/transformers/relative_mha.html). | ||||
|  | ||||
| * [GPT Architecture](https://lab-ml.com/labml_nn/transformers/gpt) | ||||
| * [kNN-LM: Generalization through Memorization](https://lab-ml.com/labml_nn/transformers/knn) | ||||
| * [Feedback Transformer](https://lab-ml.com/labml_nn/transformers/feedback) | ||||
|  | ||||
|  | ||||
| @ -18,6 +18,10 @@ and derivatives and enhancements of it. | ||||
| * [Transformer Encoder and Decoder Models](models.html) | ||||
| * [Fixed positional encoding](positional_encoding.html) | ||||
|  | ||||
| ## [GPT Architecture](gpt) | ||||
|  | ||||
| This is an implementation of GPT-2 architecture. | ||||
|  | ||||
| ## [kNN-LM](knn) | ||||
|  | ||||
| This is an implementation of the paper | ||||
|  | ||||
| @ -27,6 +27,7 @@ contains implementations for | ||||
| and | ||||
| [relative multi-headed attention](https://lab-ml.com/labml_nn/transformers/relative_mha.html). | ||||
|  | ||||
| * [GPT Architecture](https://lab-ml.com/labml_nn/transformers/gpt) | ||||
| * [kNN-LM: Generalization through Memorization](https://lab-ml.com/labml_nn/transformers/knn) | ||||
| * [Feedback Transformer](https://lab-ml.com/labml_nn/transformers/feedback) | ||||
|  | ||||
|  | ||||
		Reference in New Issue
	
	Block a user
	 Varuna Jayasiri
					Varuna Jayasiri