mirror of
https://github.com/yunjey/pytorch-tutorial.git
synced 2025-07-07 18:14:17 +08:00
Merge pull request #143 from AutuanLiu/patch01
clip_grad_norm is now deprecated
This commit is contained in:
@ -3,7 +3,7 @@
|
|||||||
import torch
|
import torch
|
||||||
import torch.nn as nn
|
import torch.nn as nn
|
||||||
import numpy as np
|
import numpy as np
|
||||||
from torch.nn.utils import clip_grad_norm
|
from torch.nn.utils import clip_grad_norm_
|
||||||
from data_utils import Dictionary, Corpus
|
from data_utils import Dictionary, Corpus
|
||||||
|
|
||||||
|
|
||||||
@ -78,7 +78,7 @@ for epoch in range(num_epochs):
|
|||||||
# Backward and optimize
|
# Backward and optimize
|
||||||
model.zero_grad()
|
model.zero_grad()
|
||||||
loss.backward()
|
loss.backward()
|
||||||
clip_grad_norm(model.parameters(), 0.5)
|
clip_grad_norm_(model.parameters(), 0.5)
|
||||||
optimizer.step()
|
optimizer.step()
|
||||||
|
|
||||||
step = (i+1) // seq_length
|
step = (i+1) // seq_length
|
||||||
|
Reference in New Issue
Block a user