From 066a0429e57330ae1cd19e9657dfb798ad38080a Mon Sep 17 00:00:00 2001 From: LYFxiaoan Date: Tue, 15 Jul 2025 11:55:39 +0800 Subject: [PATCH] =?UTF-8?q?=E6=95=B0=E6=8D=AE=E8=AE=AD=E7=BB=83=E6=A8=A1?= =?UTF-8?q?=E5=9E=8B=E9=80=9A=E8=BF=87?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- server/trainers/mlstm_trainer.py | 2 +- server/trainers/transformer_trainer.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/server/trainers/mlstm_trainer.py b/server/trainers/mlstm_trainer.py index fbe1816..55e951d 100644 --- a/server/trainers/mlstm_trainer.py +++ b/server/trainers/mlstm_trainer.py @@ -340,7 +340,7 @@ def train_product_model_with_mlstm( criterion = nn.MSELoss() optimizer = optim.Adam(model.parameters(), lr=learning_rate) - scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(optimizer, 'min', patience=patience // 2, factor=0.5, verbose=True) + scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(optimizer, 'min', patience=patience // 2, factor=0.5) emit_progress("数据预处理完成,开始模型训练...", progress=10) diff --git a/server/trainers/transformer_trainer.py b/server/trainers/transformer_trainer.py index 74a612e..c902462 100644 --- a/server/trainers/transformer_trainer.py +++ b/server/trainers/transformer_trainer.py @@ -279,7 +279,7 @@ def train_product_model_with_transformer( criterion = nn.MSELoss() optimizer = optim.Adam(model.parameters(), lr=learning_rate) - scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(optimizer, 'min', patience=patience // 2, factor=0.5, verbose=True) + scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(optimizer, 'min', patience=patience // 2, factor=0.5) # 训练模型 train_losses = []