diff --git a/server/trainers/mlstm_trainer.py b/server/trainers/mlstm_trainer.py index fbe1816..55e951d 100644 --- a/server/trainers/mlstm_trainer.py +++ b/server/trainers/mlstm_trainer.py @@ -340,7 +340,7 @@ def train_product_model_with_mlstm( criterion = nn.MSELoss() optimizer = optim.Adam(model.parameters(), lr=learning_rate) - scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(optimizer, 'min', patience=patience // 2, factor=0.5, verbose=True) + scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(optimizer, 'min', patience=patience // 2, factor=0.5) emit_progress("数据预处理完成,开始模型训练...", progress=10) diff --git a/server/trainers/transformer_trainer.py b/server/trainers/transformer_trainer.py index 74a612e..c902462 100644 --- a/server/trainers/transformer_trainer.py +++ b/server/trainers/transformer_trainer.py @@ -279,7 +279,7 @@ def train_product_model_with_transformer( criterion = nn.MSELoss() optimizer = optim.Adam(model.parameters(), lr=learning_rate) - scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(optimizer, 'min', patience=patience // 2, factor=0.5, verbose=True) + scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(optimizer, 'min', patience=patience // 2, factor=0.5) # 训练模型 train_losses = []