From fbb9f21e531ec185538b54622370a87499ca853c Mon Sep 17 00:00:00 2001 From: RVC-Boss <129054828+RVC-Boss@users.noreply.github.com> Date: Sun, 23 Feb 2025 20:26:18 +0800 Subject: [PATCH] Add files via upload --- GPT_SoVITS/s2_train.py | 4 ++-- GPT_SoVITS/s2_train_v3.py | 4 ++-- GPT_SoVITS/s2_train_v3_lora.py | 4 ++-- 3 files changed, 6 insertions(+), 6 deletions(-) diff --git a/GPT_SoVITS/s2_train.py b/GPT_SoVITS/s2_train.py index 2c7a426..4d88ee8 100644 --- a/GPT_SoVITS/s2_train.py +++ b/GPT_SoVITS/s2_train.py @@ -251,8 +251,8 @@ def run(rank, n_gpus, hps): scaler = GradScaler(enabled=hps.train.fp16_run) + print("start training from epoch %s" % epoch_str) for epoch in range(epoch_str, hps.train.epochs + 1): - print("start training from epoch %s"%epoch) if rank == 0: train_and_evaluate( rank, @@ -282,7 +282,7 @@ def run(rank, n_gpus, hps): ) scheduler_g.step() scheduler_d.step() - print("training done") + print("training done") def train_and_evaluate( diff --git a/GPT_SoVITS/s2_train_v3.py b/GPT_SoVITS/s2_train_v3.py index 1d8ff30..9933dee 100644 --- a/GPT_SoVITS/s2_train_v3.py +++ b/GPT_SoVITS/s2_train_v3.py @@ -225,8 +225,8 @@ def run(rank, n_gpus, hps): scaler = GradScaler(enabled=hps.train.fp16_run) net_d=optim_d=scheduler_d=None + print("start training from epoch %s" % epoch_str) for epoch in range(epoch_str, hps.train.epochs + 1): - print("start training from epoch %s"%epoch) if rank == 0: train_and_evaluate( rank, @@ -256,7 +256,7 @@ def run(rank, n_gpus, hps): ) scheduler_g.step() # scheduler_d.step() - print("training done") + print("training done") def train_and_evaluate( diff --git a/GPT_SoVITS/s2_train_v3_lora.py b/GPT_SoVITS/s2_train_v3_lora.py index f2d5f83..75b3415 100644 --- a/GPT_SoVITS/s2_train_v3_lora.py +++ b/GPT_SoVITS/s2_train_v3_lora.py @@ -198,8 +198,8 @@ def run(rank, n_gpus, hps): scaler = GradScaler(enabled=hps.train.fp16_run) net_d=optim_d=scheduler_d=None + print("start training from epoch %s"%epoch_str) for epoch in range(epoch_str, hps.train.epochs + 1): - print("start training from epoch %s"%epoch) if rank == 0: train_and_evaluate( rank, @@ -228,7 +228,7 @@ def run(rank, n_gpus, hps): None, ) scheduler_g.step() - print("training done") + print("training done") def train_and_evaluate( rank, epoch, hps, nets, optims, schedulers, scaler, loaders, logger, writers