From b39ac9ebcd912d4168463f84c546feaa7e80881f Mon Sep 17 00:00:00 2001 From: Varuna Jayasiri Date: Wed, 17 Feb 2021 18:35:50 +0530 Subject: [PATCH] =?UTF-8?q?=F0=9F=90=9B=20typo?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- labml_nn/transformers/compressive/__init__.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/labml_nn/transformers/compressive/__init__.py b/labml_nn/transformers/compressive/__init__.py index c9a14a13..3d8edde2 100644 --- a/labml_nn/transformers/compressive/__init__.py +++ b/labml_nn/transformers/compressive/__init__.py @@ -182,7 +182,7 @@ class AttentionReconstructionLoss: c_mem = layer.compress(mem) h = self.norm(layer.norm_self_attn, h) - mem = self.norm(layer.norm_self_attn, h) + mem = self.norm(layer.norm_self_attn, mem) c_mem = self.norm(layer.norm_self_attn, c_mem) return self.loss_func(self.attn(layer.self_attn, h, mem, mem),