mirror of
https://github.com/RVC-Boss/GPT-SoVITS.git
synced 2025-04-05 04:22:46 +08:00
Add files via upload
This commit is contained in:
parent
aa07216bba
commit
fbb9f21e53
@ -251,8 +251,8 @@ def run(rank, n_gpus, hps):
|
|||||||
|
|
||||||
scaler = GradScaler(enabled=hps.train.fp16_run)
|
scaler = GradScaler(enabled=hps.train.fp16_run)
|
||||||
|
|
||||||
|
print("start training from epoch %s" % epoch_str)
|
||||||
for epoch in range(epoch_str, hps.train.epochs + 1):
|
for epoch in range(epoch_str, hps.train.epochs + 1):
|
||||||
print("start training from epoch %s"%epoch)
|
|
||||||
if rank == 0:
|
if rank == 0:
|
||||||
train_and_evaluate(
|
train_and_evaluate(
|
||||||
rank,
|
rank,
|
||||||
@ -282,7 +282,7 @@ def run(rank, n_gpus, hps):
|
|||||||
)
|
)
|
||||||
scheduler_g.step()
|
scheduler_g.step()
|
||||||
scheduler_d.step()
|
scheduler_d.step()
|
||||||
print("training done")
|
print("training done")
|
||||||
|
|
||||||
|
|
||||||
def train_and_evaluate(
|
def train_and_evaluate(
|
||||||
|
@ -225,8 +225,8 @@ def run(rank, n_gpus, hps):
|
|||||||
scaler = GradScaler(enabled=hps.train.fp16_run)
|
scaler = GradScaler(enabled=hps.train.fp16_run)
|
||||||
|
|
||||||
net_d=optim_d=scheduler_d=None
|
net_d=optim_d=scheduler_d=None
|
||||||
|
print("start training from epoch %s" % epoch_str)
|
||||||
for epoch in range(epoch_str, hps.train.epochs + 1):
|
for epoch in range(epoch_str, hps.train.epochs + 1):
|
||||||
print("start training from epoch %s"%epoch)
|
|
||||||
if rank == 0:
|
if rank == 0:
|
||||||
train_and_evaluate(
|
train_and_evaluate(
|
||||||
rank,
|
rank,
|
||||||
@ -256,7 +256,7 @@ def run(rank, n_gpus, hps):
|
|||||||
)
|
)
|
||||||
scheduler_g.step()
|
scheduler_g.step()
|
||||||
# scheduler_d.step()
|
# scheduler_d.step()
|
||||||
print("training done")
|
print("training done")
|
||||||
|
|
||||||
|
|
||||||
def train_and_evaluate(
|
def train_and_evaluate(
|
||||||
|
@ -198,8 +198,8 @@ def run(rank, n_gpus, hps):
|
|||||||
scaler = GradScaler(enabled=hps.train.fp16_run)
|
scaler = GradScaler(enabled=hps.train.fp16_run)
|
||||||
|
|
||||||
net_d=optim_d=scheduler_d=None
|
net_d=optim_d=scheduler_d=None
|
||||||
|
print("start training from epoch %s"%epoch_str)
|
||||||
for epoch in range(epoch_str, hps.train.epochs + 1):
|
for epoch in range(epoch_str, hps.train.epochs + 1):
|
||||||
print("start training from epoch %s"%epoch)
|
|
||||||
if rank == 0:
|
if rank == 0:
|
||||||
train_and_evaluate(
|
train_and_evaluate(
|
||||||
rank,
|
rank,
|
||||||
@ -228,7 +228,7 @@ def run(rank, n_gpus, hps):
|
|||||||
None,
|
None,
|
||||||
)
|
)
|
||||||
scheduler_g.step()
|
scheduler_g.step()
|
||||||
print("training done")
|
print("training done")
|
||||||
|
|
||||||
def train_and_evaluate(
|
def train_and_evaluate(
|
||||||
rank, epoch, hps, nets, optims, schedulers, scaler, loaders, logger, writers
|
rank, epoch, hps, nets, optims, schedulers, scaler, loaders, logger, writers
|
||||||
|
Loading…
x
Reference in New Issue
Block a user