数据训练模型通过
This commit is contained in:
parent
6c11aff234
commit
066a0429e5
@ -340,7 +340,7 @@ def train_product_model_with_mlstm(
|
|||||||
|
|
||||||
criterion = nn.MSELoss()
|
criterion = nn.MSELoss()
|
||||||
optimizer = optim.Adam(model.parameters(), lr=learning_rate)
|
optimizer = optim.Adam(model.parameters(), lr=learning_rate)
|
||||||
scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(optimizer, 'min', patience=patience // 2, factor=0.5, verbose=True)
|
scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(optimizer, 'min', patience=patience // 2, factor=0.5)
|
||||||
|
|
||||||
emit_progress("数据预处理完成,开始模型训练...", progress=10)
|
emit_progress("数据预处理完成,开始模型训练...", progress=10)
|
||||||
|
|
||||||
|
@ -279,7 +279,7 @@ def train_product_model_with_transformer(
|
|||||||
|
|
||||||
criterion = nn.MSELoss()
|
criterion = nn.MSELoss()
|
||||||
optimizer = optim.Adam(model.parameters(), lr=learning_rate)
|
optimizer = optim.Adam(model.parameters(), lr=learning_rate)
|
||||||
scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(optimizer, 'min', patience=patience // 2, factor=0.5, verbose=True)
|
scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(optimizer, 'min', patience=patience // 2, factor=0.5)
|
||||||
|
|
||||||
# 训练模型
|
# 训练模型
|
||||||
train_losses = []
|
train_losses = []
|
||||||
|
Loading…
x
Reference in New Issue
Block a user