From 82e04d7815a8cf1935f2fec5cddc03bdb87c8484 Mon Sep 17 00:00:00 2001 From: tianhao zhang <15600919271@163.com> Date: Thu, 8 Sep 2022 12:28:19 +0000 Subject: [PATCH] fix trianer --- examples/aishell/asr1/run.sh | 4 ++-- paddlespeech/s2t/training/trainer.py | 5 ++++- 2 files changed, 6 insertions(+), 3 deletions(-) diff --git a/examples/aishell/asr1/run.sh b/examples/aishell/asr1/run.sh index bd4f50e3f..701dcd2ac 100644 --- a/examples/aishell/asr1/run.sh +++ b/examples/aishell/asr1/run.sh @@ -2,8 +2,8 @@ source path.sh set -e -gpus=0,1,2,3 -stage=0 +gpus=1 +stage=1 stop_stage=50 conf_path=conf/conformer.yaml ips= #xx.xx.xx.xx,xx.xx.xx.xx diff --git a/paddlespeech/s2t/training/trainer.py b/paddlespeech/s2t/training/trainer.py index a7eb9892d..d1bd30fef 100644 --- a/paddlespeech/s2t/training/trainer.py +++ b/paddlespeech/s2t/training/trainer.py @@ -19,6 +19,9 @@ from pathlib import Path import paddle from paddle import distributed as dist +import pdb +pdb.set_trace() +dist.init_parallel_env() from visualdl import LogWriter from paddlespeech.s2t.training.reporter import ObsScope @@ -176,7 +179,7 @@ class Trainer(): def init_parallel(self): """Init environment for multiprocess training. """ - dist.init_parallel_env() + # dist.init_parallel_env() @mp_tools.rank_zero_only def save(self, tag=None, infos: dict=None):