From a16173ffc4fcf0474d2d32d1cf8c9cf124afe34d Mon Sep 17 00:00:00 2001 From: e Date: Fri, 9 Nov 2018 16:21:42 +0800 Subject: [PATCH] clip_grad_norm is now deprecated --- tutorials/02-intermediate/language_model/main.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/tutorials/02-intermediate/language_model/main.py b/tutorials/02-intermediate/language_model/main.py index c7555e7..3c03db0 100644 --- a/tutorials/02-intermediate/language_model/main.py +++ b/tutorials/02-intermediate/language_model/main.py @@ -3,7 +3,7 @@ import torch import torch.nn as nn import numpy as np -from torch.nn.utils import clip_grad_norm +from torch.nn.utils import clip_grad_norm_ from data_utils import Dictionary, Corpus @@ -78,7 +78,7 @@ for epoch in range(num_epochs): # Backward and optimize model.zero_grad() loss.backward() - clip_grad_norm(model.parameters(), 0.5) + clip_grad_norm_(model.parameters(), 0.5) optimizer.step() step = (i+1) // seq_length