修改gradient accumulation下loss计算逻辑
This commit is contained in:
2
train.py
2
train.py
@@ -221,7 +221,7 @@ def main():
|
||||
step + 1,
|
||||
piece_num,
|
||||
epoch + 1,
|
||||
running_loss * gradient_accumulation / log_step))
|
||||
running_loss / log_step))
|
||||
running_loss = 0
|
||||
piece_num += 1
|
||||
|
||||
|
||||
@@ -197,7 +197,7 @@ def main():
|
||||
(step + 1) // gradient_accumulation,
|
||||
piece_num,
|
||||
epoch + 1,
|
||||
running_loss * gradient_accumulation / log_step))
|
||||
running_loss / log_step))
|
||||
running_loss = 0
|
||||
piece_num += 1
|
||||
|
||||
|
||||
Reference in New Issue
Block a user