|
|
|
# Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved.
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
from typing import Any
|
|
|
|
from typing import Dict
|
|
|
|
from typing import List
|
|
|
|
from typing import Text
|
|
|
|
|
|
|
|
import jsonlines
|
|
|
|
import numpy as np
|
|
|
|
import paddle
|
|
|
|
from paddle.io import BatchSampler
|
|
|
|
from paddle.io import DataLoader
|
|
|
|
from paddle.io import DistributedBatchSampler
|
|
|
|
|
|
|
|
from paddlespeech.s2t.io.batchfy import make_batchset
|
|
|
|
from paddlespeech.s2t.io.converter import CustomConverter
|
|
|
|
from paddlespeech.s2t.io.dataset import TransformDataset
|
|
|
|
from paddlespeech.s2t.io.reader import LoadInputsAndTargets
|
|
|
|
from paddlespeech.s2t.utils.log import Log
|
|
|
|
|
|
|
|
import paddlespeech.audio.streamdata as streamdata
|
|
|
|
from paddlespeech.audio.text.text_featurizer import TextFeaturizer
|
|
|
|
from yacs.config import CfgNode
|
|
|
|
|
|
|
|
__all__ = ["BatchDataLoader", "StreamDataLoader"]
|
|
|
|
|
|
|
|
logger = Log(__name__).getlog()
|
|
|
|
|
|
|
|
|
|
|
|
def feat_dim_and_vocab_size(data_json: List[Dict[Text, Any]],
|
|
|
|
mode: Text="asr",
|
|
|
|
iaxis=0,
|
|
|
|
oaxis=0):
|
|
|
|
if mode == 'asr':
|
|
|
|
feat_dim = data_json[0]['input'][oaxis]['shape'][1]
|
|
|
|
vocab_size = data_json[0]['output'][oaxis]['shape'][1]
|
|
|
|
else:
|
|
|
|
raise ValueError(f"{mode} mode not support!")
|
|
|
|
return feat_dim, vocab_size
|
|
|
|
|
|
|
|
|
|
|
|
def batch_collate(x):
|
|
|
|
"""de-minibatch, since user compose batch.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
x (List[Tuple]): [(utts, xs, ilens, ys, olens)]
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
Tuple: (utts, xs, ilens, ys, olens)
|
|
|
|
"""
|
|
|
|
return x[0]
|
|
|
|
|
|
|
|
def read_preprocess_cfg(preprocess_conf_file):
|
|
|
|
augment_conf = dict()
|
|
|
|
preprocess_cfg = CfgNode(new_allowed=True)
|
|
|
|
preprocess_cfg.merge_from_file(preprocess_conf_file)
|
|
|
|
for idx, process in enumerate(preprocess_cfg["process"]):
|
|
|
|
opts = dict(process)
|
|
|
|
process_type = opts.pop("type")
|
|
|
|
if process_type == 'time_warp':
|
|
|
|
augment_conf['max_w'] = process['max_time_warp']
|
|
|
|
augment_conf['w_inplace'] = process['inplace']
|
|
|
|
augment_conf['w_mode'] = process['mode']
|
|
|
|
if process_type == 'freq_mask':
|
|
|
|
augment_conf['max_f'] = process['F']
|
|
|
|
augment_conf['num_f_mask'] = process['n_mask']
|
|
|
|
augment_conf['f_inplace'] = process['inplace']
|
|
|
|
augment_conf['f_replace_with_zero'] = process['replace_with_zero']
|
|
|
|
if process_type == 'time_mask':
|
|
|
|
augment_conf['max_t'] = process['T']
|
|
|
|
augment_conf['num_t_mask'] = process['n_mask']
|
|
|
|
augment_conf['t_inplace'] = process['inplace']
|
|
|
|
augment_conf['t_replace_with_zero'] = process['replace_with_zero']
|
|
|
|
return augment_conf
|
|
|
|
|
|
|
|
class StreamDataLoader():
|
|
|
|
def __init__(self,
|
|
|
|
manifest_file: str,
|
|
|
|
train_mode: bool,
|
|
|
|
unit_type: str='char',
|
|
|
|
batch_size: int=0,
|
|
|
|
preprocess_conf=None,
|
|
|
|
num_mel_bins=80,
|
|
|
|
frame_length=25,
|
|
|
|
frame_shift=10,
|
|
|
|
dither=0.0,
|
|
|
|
minlen_in: float=0.0,
|
|
|
|
maxlen_in: float=float('inf'),
|
|
|
|
minlen_out: float=0.0,
|
|
|
|
maxlen_out: float=float('inf'),
|
|
|
|
resample_rate: int=16000,
|
|
|
|
shuffle_size: int=10000,
|
|
|
|
sort_size: int=1000,
|
|
|
|
n_iter_processes: int=1,
|
|
|
|
prefetch_factor: int=2,
|
|
|
|
dist_sampler: bool=False,
|
|
|
|
cmvn_file="data/mean_std.json",
|
|
|
|
vocab_filepath='data/lang_char/vocab.txt'):
|
|
|
|
self.manifest_file = manifest_file
|
|
|
|
self.train_model = train_mode
|
|
|
|
self.batch_size = batch_size
|
|
|
|
self.prefetch_factor = prefetch_factor
|
|
|
|
self.dist_sampler = dist_sampler
|
|
|
|
self.n_iter_processes = n_iter_processes
|
|
|
|
|
|
|
|
text_featurizer = TextFeaturizer(unit_type, vocab_filepath)
|
|
|
|
symbol_table = text_featurizer.vocab_dict
|
|
|
|
self.feat_dim = num_mel_bins
|
|
|
|
self.vocab_size = text_featurizer.vocab_size
|
|
|
|
|
|
|
|
augment_conf = read_preprocess_cfg(preprocess_conf)
|
|
|
|
|
|
|
|
# The list of shard
|
|
|
|
shardlist = []
|
|
|
|
with open(manifest_file, "r") as f:
|
|
|
|
for line in f.readlines():
|
|
|
|
shardlist.append(line.strip())
|
|
|
|
world_size = 1
|
|
|
|
try:
|
|
|
|
world_size = paddle.distributed.get_world_size()
|
|
|
|
except Exception as e:
|
|
|
|
logger.warninig(e)
|
|
|
|
logger.warninig("can not get world_size using paddle.distributed.get_world_size(), use world_size=1")
|
|
|
|
assert(len(shardlist) >= world_size, "the length of shard list should >= number of gpus/xpus/...")
|
|
|
|
|
|
|
|
update_n_iter_processes = int(max(min(len(shardlist)/world_size - 1, self.n_iter_processes), 0))
|
|
|
|
logger.info(f"update_n_iter_processes {update_n_iter_processes}")
|
|
|
|
if update_n_iter_processes != self.n_iter_processes:
|
|
|
|
self.n_iter_processes = update_n_iter_processes
|
|
|
|
logger.info(f"change nun_workers to {self.n_iter_processes}")
|
|
|
|
|
|
|
|
if self.dist_sampler:
|
|
|
|
base_dataset = streamdata.DataPipeline(
|
|
|
|
streamdata.SimpleShardList(shardlist),
|
|
|
|
streamdata.split_by_node if train_mode else streamdata.placeholder(),
|
|
|
|
streamdata.split_by_worker,
|
|
|
|
streamdata.tarfile_to_samples(streamdata.reraise_exception)
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
base_dataset = streamdata.DataPipeline(
|
|
|
|
streamdata.SimpleShardList(shardlist),
|
|
|
|
streamdata.split_by_worker,
|
|
|
|
streamdata.tarfile_to_samples(streamdata.reraise_exception)
|
|
|
|
)
|
|
|
|
|
|
|
|
self.dataset = base_dataset.append_list(
|
|
|
|
streamdata.audio_tokenize(symbol_table),
|
|
|
|
streamdata.audio_data_filter(frame_shift=frame_shift, max_length=maxlen_in, min_length=minlen_in, token_max_length=maxlen_out, token_min_length=minlen_out),
|
|
|
|
streamdata.audio_resample(resample_rate=resample_rate),
|
|
|
|
streamdata.audio_compute_fbank(num_mel_bins=num_mel_bins, frame_length=frame_length, frame_shift=frame_shift, dither=dither),
|
|
|
|
streamdata.audio_spec_aug(**augment_conf) if train_mode else streamdata.placeholder(), # num_t_mask=2, num_f_mask=2, max_t=40, max_f=30, max_w=80)
|
|
|
|
streamdata.shuffle(shuffle_size),
|
|
|
|
streamdata.sort(sort_size=sort_size),
|
|
|
|
streamdata.batched(batch_size),
|
|
|
|
streamdata.audio_padding(),
|
|
|
|
streamdata.audio_cmvn(cmvn_file)
|
|
|
|
)
|
|
|
|
|
|
|
|
if paddle.__version__ >= '2.3.2':
|
|
|
|
self.loader = streamdata.WebLoader(
|
|
|
|
self.dataset,
|
|
|
|
num_workers=self.n_iter_processes,
|
|
|
|
prefetch_factor = self.prefetch_factor,
|
|
|
|
batch_size=None
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
self.loader = streamdata.WebLoader(
|
|
|
|
self.dataset,
|
|
|
|
num_workers=self.n_iter_processes,
|
|
|
|
batch_size=None
|
|
|
|
)
|
|
|
|
|
|
|
|
def __iter__(self):
|
|
|
|
return self.loader.__iter__()
|
|
|
|
|
|
|
|
def __call__(self):
|
|
|
|
return self.__iter__()
|
|
|
|
|
|
|
|
def __len__(self):
|
|
|
|
logger.info("Stream dataloader does not support calculate the length of the dataset")
|
|
|
|
return -1
|
|
|
|
|
|
|
|
|
|
|
|
class BatchDataLoader():
|
|
|
|
def __init__(self,
|
|
|
|
json_file: str,
|
|
|
|
train_mode: bool,
|
|
|
|
sortagrad: int=0,
|
|
|
|
batch_size: int=0,
|
|
|
|
maxlen_in: float=float('inf'),
|
|
|
|
maxlen_out: float=float('inf'),
|
|
|
|
minibatches: int=0,
|
|
|
|
mini_batch_size: int=1,
|
|
|
|
batch_count: str='auto',
|
|
|
|
batch_bins: int=0,
|
|
|
|
batch_frames_in: int=0,
|
|
|
|
batch_frames_out: int=0,
|
|
|
|
batch_frames_inout: int=0,
|
|
|
|
preprocess_conf=None,
|
|
|
|
n_iter_processes: int=1,
|
|
|
|
subsampling_factor: int=1,
|
|
|
|
load_aux_input: bool=False,
|
|
|
|
load_aux_output: bool=False,
|
|
|
|
num_encs: int=1,
|
|
|
|
dist_sampler: bool=False,
|
|
|
|
shortest_first: bool=False):
|
|
|
|
self.json_file = json_file
|
|
|
|
self.train_mode = train_mode
|
|
|
|
self.use_sortagrad = sortagrad == -1 or sortagrad > 0
|
|
|
|
self.batch_size = batch_size
|
|
|
|
self.maxlen_in = maxlen_in
|
|
|
|
self.maxlen_out = maxlen_out
|
|
|
|
self.batch_count = batch_count
|
|
|
|
self.batch_bins = batch_bins
|
|
|
|
self.batch_frames_in = batch_frames_in
|
|
|
|
self.batch_frames_out = batch_frames_out
|
|
|
|
self.batch_frames_inout = batch_frames_inout
|
|
|
|
self.subsampling_factor = subsampling_factor
|
|
|
|
self.num_encs = num_encs
|
|
|
|
self.preprocess_conf = preprocess_conf
|
|
|
|
self.n_iter_processes = n_iter_processes
|
|
|
|
self.load_aux_input = load_aux_input
|
|
|
|
self.load_aux_output = load_aux_output
|
|
|
|
self.dist_sampler = dist_sampler
|
|
|
|
self.shortest_first = shortest_first
|
|
|
|
|
|
|
|
# read json data
|
|
|
|
with jsonlines.open(json_file, 'r') as reader:
|
|
|
|
self.data_json = list(reader)
|
|
|
|
|
|
|
|
self.feat_dim, self.vocab_size = feat_dim_and_vocab_size(
|
|
|
|
self.data_json, mode='asr')
|
|
|
|
|
|
|
|
# make minibatch list (variable length)
|
|
|
|
self.minibaches = make_batchset(
|
|
|
|
self.data_json,
|
|
|
|
batch_size,
|
|
|
|
maxlen_in,
|
|
|
|
maxlen_out,
|
|
|
|
minibatches, # for debug
|
|
|
|
min_batch_size=mini_batch_size,
|
|
|
|
shortest_first=self.shortest_first or self.use_sortagrad,
|
|
|
|
count=batch_count,
|
|
|
|
batch_bins=batch_bins,
|
|
|
|
batch_frames_in=batch_frames_in,
|
|
|
|
batch_frames_out=batch_frames_out,
|
|
|
|
batch_frames_inout=batch_frames_inout,
|
|
|
|
iaxis=0,
|
|
|
|
oaxis=0, )
|
|
|
|
|
|
|
|
# data reader
|
|
|
|
self.reader = LoadInputsAndTargets(
|
|
|
|
mode="asr",
|
|
|
|
load_output=True,
|
|
|
|
preprocess_conf=preprocess_conf,
|
|
|
|
preprocess_args={"train":
|
|
|
|
train_mode}, # Switch the mode of preprocessing
|
|
|
|
)
|
|
|
|
|
|
|
|
# Setup a converter
|
|
|
|
if num_encs == 1:
|
|
|
|
self.converter = CustomConverter(
|
|
|
|
subsampling_factor=subsampling_factor,
|
|
|
|
dtype=np.float32,
|
|
|
|
load_aux_input=load_aux_input,
|
|
|
|
load_aux_output=load_aux_output)
|
|
|
|
else:
|
|
|
|
assert NotImplementedError("not impl CustomConverterMulEnc.")
|
|
|
|
|
|
|
|
# hack to make batchsize argument as 1
|
|
|
|
# actual bathsize is included in a list
|
|
|
|
# default collate function converts numpy array to paddle tensor
|
|
|
|
# we used an empty collate function instead which returns list
|
|
|
|
self.dataset = TransformDataset(self.minibaches, self.converter,
|
|
|
|
self.reader)
|
|
|
|
|
|
|
|
if self.dist_sampler:
|
|
|
|
self.batch_sampler = DistributedBatchSampler(
|
|
|
|
dataset=self.dataset,
|
|
|
|
batch_size=1,
|
|
|
|
shuffle=not self.use_sortagrad if self.train_mode else False,
|
|
|
|
drop_last=False, )
|
|
|
|
else:
|
|
|
|
self.batch_sampler = BatchSampler(
|
|
|
|
dataset=self.dataset,
|
|
|
|
batch_size=1,
|
|
|
|
shuffle=not self.use_sortagrad if self.train_mode else False,
|
|
|
|
drop_last=False, )
|
|
|
|
|
|
|
|
self.dataloader = DataLoader(
|
|
|
|
dataset=self.dataset,
|
|
|
|
batch_sampler=self.batch_sampler,
|
|
|
|
collate_fn=batch_collate,
|
|
|
|
num_workers=self.n_iter_processes, )
|
|
|
|
|
|
|
|
def __len__(self):
|
|
|
|
return len(self.dataloader)
|
|
|
|
|
|
|
|
def __iter__(self):
|
|
|
|
return self.dataloader.__iter__()
|
|
|
|
|
|
|
|
def __call__(self):
|
|
|
|
return self.__iter__()
|
|
|
|
|
|
|
|
def __repr__(self):
|
|
|
|
echo = f"<{self.__class__.__module__}.{self.__class__.__name__} object at {hex(id(self))}> "
|
|
|
|
echo += f"train_mode: {self.train_mode}, "
|
|
|
|
echo += f"sortagrad: {self.use_sortagrad}, "
|
|
|
|
echo += f"batch_size: {self.batch_size}, "
|
|
|
|
echo += f"maxlen_in: {self.maxlen_in}, "
|
|
|
|
echo += f"maxlen_out: {self.maxlen_out}, "
|
|
|
|
echo += f"batch_count: {self.batch_count}, "
|
|
|
|
echo += f"batch_bins: {self.batch_bins}, "
|
|
|
|
echo += f"batch_frames_in: {self.batch_frames_in}, "
|
|
|
|
echo += f"batch_frames_out: {self.batch_frames_out}, "
|
|
|
|
echo += f"batch_frames_inout: {self.batch_frames_inout}, "
|
|
|
|
echo += f"subsampling_factor: {self.subsampling_factor}, "
|
|
|
|
echo += f"num_encs: {self.num_encs}, "
|
|
|
|
echo += f"num_workers: {self.n_iter_processes}, "
|
|
|
|
echo += f"load_aux_input: {self.load_aux_input}, "
|
|
|
|
echo += f"load_aux_output: {self.load_aux_output}, "
|
|
|
|
echo += f"dist_sampler: {self.dist_sampler}, "
|
|
|
|
echo += f"shortest_first: {self.shortest_first}, "
|
|
|
|
echo += f"file: {self.json_file}"
|
|
|
|
return echo
|
|
|
|
|
|
|
|
|
|
|
|
class DataLoaderFactory():
|
|
|
|
@staticmethod
|
|
|
|
def get_dataloader(mode: str, config, args):
|
|
|
|
config = config.clone()
|
|
|
|
use_streamdata = config.get("use_stream_data", False)
|
|
|
|
if use_streamdata:
|
|
|
|
if mode == 'train':
|
|
|
|
config['manifest'] = config.train_manifest
|
|
|
|
config['train_mode'] = True
|
|
|
|
elif mode == 'valid':
|
|
|
|
config['manifest'] = config.dev_manifest
|
|
|
|
config['train_mode'] = False
|
|
|
|
elif model == 'test' or mode == 'align':
|
|
|
|
config['manifest'] = config.test_manifest
|
|
|
|
config['train_mode'] = False
|
|
|
|
config['dither'] = 0.0
|
|
|
|
config['minlen_in'] = 0.0
|
|
|
|
config['maxlen_in'] = float('inf')
|
|
|
|
config['minlen_out'] = 0
|
|
|
|
config['maxlen_out'] = float('inf')
|
|
|
|
config['dist_sampler'] = False
|
|
|
|
else:
|
|
|
|
raise KeyError("not valid mode type!!, please input one of 'train, valid, test, align'")
|
|
|
|
return StreamDataLoader(
|
|
|
|
manifest_file=config.manifest,
|
|
|
|
train_mode=config.train_mode,
|
|
|
|
unit_type=config.unit_type,
|
|
|
|
preprocess_conf=config.preprocess_config,
|
|
|
|
batch_size=config.batch_size,
|
|
|
|
num_mel_bins=config.feat_dim,
|
|
|
|
frame_length=config.window_ms,
|
|
|
|
frame_shift=config.stride_ms,
|
|
|
|
dither=config.dither,
|
|
|
|
minlen_in=config.minlen_in,
|
|
|
|
maxlen_in=config.maxlen_in,
|
|
|
|
minlen_out=config.minlen_out,
|
|
|
|
maxlen_out=config.maxlen_out,
|
|
|
|
resample_rate=config.resample_rate,
|
|
|
|
shuffle_size=config.shuffle_size,
|
|
|
|
sort_size=config.sort_size,
|
|
|
|
n_iter_processes=config.num_workers,
|
|
|
|
prefetch_factor=config.prefetch_factor,
|
|
|
|
dist_sampler=config.dist_sampler,
|
|
|
|
cmvn_file=config.cmvn_file,
|
|
|
|
vocab_filepath=config.vocab_filepath,
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
if mode == 'train':
|
|
|
|
config['manifest'] = config.train_manifest
|
|
|
|
config['train_mode'] = True
|
|
|
|
config['mini_batch_size'] = args.ngpu
|
|
|
|
config['subsampling_factor'] = 1
|
|
|
|
config['num_encs'] = 1
|
|
|
|
config['shortest_first'] = False
|
|
|
|
elif mode == 'valid':
|
|
|
|
config['manifest'] = config.dev_manifest
|
|
|
|
config['train_mode'] = False
|
|
|
|
config['sortagrad'] = False
|
|
|
|
config['maxlen_in'] = float('inf')
|
|
|
|
config['maxlen_out'] = float('inf')
|
|
|
|
config['minibatches'] = 0
|
|
|
|
config['mini_batch_size'] = args.ngpu
|
|
|
|
config['batch_count'] = 'auto'
|
|
|
|
config['batch_bins'] = 0
|
|
|
|
config['batch_frames_in'] = 0
|
|
|
|
config['batch_frames_out'] = 0
|
|
|
|
config['batch_frames_inout'] = 0
|
|
|
|
config['subsampling_factor'] = 1
|
|
|
|
config['num_encs'] = 1
|
|
|
|
config['shortest_first'] = False
|
|
|
|
elif mode == 'test' or mode == 'align':
|
|
|
|
config['manifest'] = config.test_manifest
|
|
|
|
config['train_mode'] = False
|
|
|
|
config['sortagrad'] = False
|
|
|
|
config['batch_size'] = config.get('decode', dict()).get(
|
|
|
|
'decode_batch_size', 1)
|
|
|
|
config['maxlen_in'] = float('inf')
|
|
|
|
config['maxlen_out'] = float('inf')
|
|
|
|
config['minibatches'] = 0
|
|
|
|
config['mini_batch_size'] = 1
|
|
|
|
config['batch_count'] = 'auto'
|
|
|
|
config['batch_bins'] = 0
|
|
|
|
config['batch_frames_in'] = 0
|
|
|
|
config['batch_frames_out'] = 0
|
|
|
|
config['batch_frames_inout'] = 0
|
|
|
|
config['num_workers'] = 1
|
|
|
|
config['subsampling_factor'] = 1
|
|
|
|
config['num_encs'] = 1
|
|
|
|
config['dist_sampler'] = False
|
|
|
|
config['shortest_first'] = False
|
|
|
|
else:
|
|
|
|
raise KeyError("not valid mode type!!, please input one of 'train, valid, test, align'")
|
|
|
|
|
|
|
|
return BatchDataLoader(
|
|
|
|
json_file=config.manifest,
|
|
|
|
train_mode=config.train_mode,
|
|
|
|
sortagrad=config.sortagrad,
|
|
|
|
batch_size=config.batch_size,
|
|
|
|
maxlen_in=config.maxlen_in,
|
|
|
|
maxlen_out=config.maxlen_out,
|
|
|
|
minibatches=config.minibatches,
|
|
|
|
mini_batch_size=config.mini_batch_size,
|
|
|
|
batch_count=config.batch_count,
|
|
|
|
batch_bins=config.batch_bins,
|
|
|
|
batch_frames_in=config.batch_frames_in,
|
|
|
|
batch_frames_out=config.batch_frames_out,
|
|
|
|
batch_frames_inout=config.batch_frames_inout,
|
|
|
|
preprocess_conf=config.preprocess_config,
|
|
|
|
n_iter_processes=config.num_workers,
|
|
|
|
subsampling_factor=config.subsampling_factor,
|
|
|
|
load_aux_output=config.get('load_transcript', None),
|
|
|
|
num_encs=config.num_encs,
|
|
|
|
dist_sampler=config.dist_sampler,
|
|
|
|
shortest_first=config.shortest_first)
|
|
|
|
|