修改gradient accumulation下loss计算逻辑
This commit is contained in:
2
train.py
2
train.py
@@ -221,7 +221,7 @@ def main():
|
|||||||
step + 1,
|
step + 1,
|
||||||
piece_num,
|
piece_num,
|
||||||
epoch + 1,
|
epoch + 1,
|
||||||
running_loss * gradient_accumulation / log_step))
|
running_loss / log_step))
|
||||||
running_loss = 0
|
running_loss = 0
|
||||||
piece_num += 1
|
piece_num += 1
|
||||||
|
|
||||||
|
|||||||
@@ -197,7 +197,7 @@ def main():
|
|||||||
(step + 1) // gradient_accumulation,
|
(step + 1) // gradient_accumulation,
|
||||||
piece_num,
|
piece_num,
|
||||||
epoch + 1,
|
epoch + 1,
|
||||||
running_loss * gradient_accumulation / log_step))
|
running_loss / log_step))
|
||||||
running_loss = 0
|
running_loss = 0
|
||||||
piece_num += 1
|
piece_num += 1
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user