|
|
|
@ -12,29 +12,17 @@
|
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
|
# limitations under the License.
|
|
|
|
|
import base64
|
|
|
|
|
import io
|
|
|
|
|
import time
|
|
|
|
|
|
|
|
|
|
import librosa
|
|
|
|
|
import numpy as np
|
|
|
|
|
import paddle
|
|
|
|
|
import soundfile as sf
|
|
|
|
|
from scipy.io import wavfile
|
|
|
|
|
|
|
|
|
|
from paddlespeech.cli.log import logger
|
|
|
|
|
from paddlespeech.cli.tts.infer import TTSExecutor
|
|
|
|
|
from paddlespeech.server.engine.base_engine import BaseEngine
|
|
|
|
|
from paddlespeech.server.utils.audio_process import change_speed
|
|
|
|
|
from paddlespeech.server.utils.errors import ErrorCode
|
|
|
|
|
from paddlespeech.server.utils.exception import ServerBaseException
|
|
|
|
|
from paddlespeech.server.utils.audio_process import float2pcm
|
|
|
|
|
from paddlespeech.server.utils.config import get_config
|
|
|
|
|
from paddlespeech.server.utils.util import denorm
|
|
|
|
|
from paddlespeech.server.utils.util import get_chunks
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import math
|
|
|
|
|
|
|
|
|
|
__all__ = ['TTSEngine']
|
|
|
|
|
|
|
|
|
|
|
|
|
|
@ -44,15 +32,16 @@ class TTSServerExecutor(TTSExecutor):
|
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
@paddle.no_grad()
|
|
|
|
|
def infer(self,
|
|
|
|
|
text: str,
|
|
|
|
|
lang: str='zh',
|
|
|
|
|
am: str='fastspeech2_csmsc',
|
|
|
|
|
spk_id: int=0,
|
|
|
|
|
am_block: int=42,
|
|
|
|
|
am_pad: int=12,
|
|
|
|
|
voc_block: int=14,
|
|
|
|
|
voc_pad: int=14,):
|
|
|
|
|
def infer(
|
|
|
|
|
self,
|
|
|
|
|
text: str,
|
|
|
|
|
lang: str='zh',
|
|
|
|
|
am: str='fastspeech2_csmsc',
|
|
|
|
|
spk_id: int=0,
|
|
|
|
|
am_block: int=42,
|
|
|
|
|
am_pad: int=12,
|
|
|
|
|
voc_block: int=14,
|
|
|
|
|
voc_pad: int=14, ):
|
|
|
|
|
"""
|
|
|
|
|
Model inference and result stored in self.output.
|
|
|
|
|
"""
|
|
|
|
@ -61,8 +50,6 @@ class TTSServerExecutor(TTSExecutor):
|
|
|
|
|
get_tone_ids = False
|
|
|
|
|
merge_sentences = False
|
|
|
|
|
frontend_st = time.time()
|
|
|
|
|
if am_name == 'speedyspeech':
|
|
|
|
|
get_tone_ids = True
|
|
|
|
|
if lang == 'zh':
|
|
|
|
|
input_ids = self.frontend.get_input_ids(
|
|
|
|
|
text,
|
|
|
|
@ -95,7 +82,7 @@ class TTSServerExecutor(TTSExecutor):
|
|
|
|
|
else:
|
|
|
|
|
mel = self.am_inference(part_phone_ids)
|
|
|
|
|
am_et = time.time()
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
# voc streaming
|
|
|
|
|
voc_upsample = self.voc_config.n_shift
|
|
|
|
|
mel_chunks = get_chunks(mel, voc_block, voc_pad, "voc")
|
|
|
|
@ -103,17 +90,19 @@ class TTSServerExecutor(TTSExecutor):
|
|
|
|
|
voc_st = time.time()
|
|
|
|
|
for i, mel_chunk in enumerate(mel_chunks):
|
|
|
|
|
sub_wav = self.voc_inference(mel_chunk)
|
|
|
|
|
front_pad = min(i*voc_block, voc_pad)
|
|
|
|
|
front_pad = min(i * voc_block, voc_pad)
|
|
|
|
|
|
|
|
|
|
if i == 0:
|
|
|
|
|
sub_wav = sub_wav[: voc_block * voc_upsample]
|
|
|
|
|
sub_wav = sub_wav[:voc_block * voc_upsample]
|
|
|
|
|
elif i == chunk_num - 1:
|
|
|
|
|
sub_wav = sub_wav[front_pad * voc_upsample : ]
|
|
|
|
|
sub_wav = sub_wav[front_pad * voc_upsample:]
|
|
|
|
|
else:
|
|
|
|
|
sub_wav = sub_wav[front_pad * voc_upsample: (front_pad + voc_block) * voc_upsample]
|
|
|
|
|
|
|
|
|
|
sub_wav = sub_wav[front_pad * voc_upsample:(
|
|
|
|
|
front_pad + voc_block) * voc_upsample]
|
|
|
|
|
|
|
|
|
|
yield sub_wav
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
class TTSEngine(BaseEngine):
|
|
|
|
|
"""TTS server engine
|
|
|
|
|
|
|
|
|
@ -128,9 +117,11 @@ class TTSEngine(BaseEngine):
|
|
|
|
|
|
|
|
|
|
def init(self, config: dict) -> bool:
|
|
|
|
|
self.executor = TTSServerExecutor()
|
|
|
|
|
|
|
|
|
|
self.config = config
|
|
|
|
|
assert "fastspeech2_csmsc" in config.am and (
|
|
|
|
|
config.voc == "hifigan_csmsc-zh" or config.voc == "mb_melgan_csmsc"
|
|
|
|
|
), 'Please check config, am support: fastspeech2, voc support: hifigan_csmsc-zh or mb_melgan_csmsc.'
|
|
|
|
|
try:
|
|
|
|
|
self.config = config
|
|
|
|
|
if self.config.device:
|
|
|
|
|
self.device = self.config.device
|
|
|
|
|
else:
|
|
|
|
@ -176,86 +167,11 @@ class TTSEngine(BaseEngine):
|
|
|
|
|
def preprocess(self, text_bese64: str=None, text_bytes: bytes=None):
|
|
|
|
|
# Convert byte to text
|
|
|
|
|
if text_bese64:
|
|
|
|
|
text_bytes = base64.b64decode(text_bese64) # base64 to bytes
|
|
|
|
|
text = text_bytes.decode('utf-8') # bytes to text
|
|
|
|
|
text_bytes = base64.b64decode(text_bese64) # base64 to bytes
|
|
|
|
|
text = text_bytes.decode('utf-8') # bytes to text
|
|
|
|
|
|
|
|
|
|
return text
|
|
|
|
|
|
|
|
|
|
def postprocess(self,
|
|
|
|
|
wav,
|
|
|
|
|
original_fs: int,
|
|
|
|
|
target_fs: int=0,
|
|
|
|
|
volume: float=1.0,
|
|
|
|
|
speed: float=1.0,
|
|
|
|
|
audio_path: str=None):
|
|
|
|
|
"""Post-processing operations, including speech, volume, sample rate, save audio file
|
|
|
|
|
|
|
|
|
|
Args:
|
|
|
|
|
wav (numpy(float)): Synthesized audio sample points
|
|
|
|
|
original_fs (int): original audio sample rate
|
|
|
|
|
target_fs (int): target audio sample rate
|
|
|
|
|
volume (float): target volume
|
|
|
|
|
speed (float): target speed
|
|
|
|
|
|
|
|
|
|
Raises:
|
|
|
|
|
ServerBaseException: Throws an exception if the change speed unsuccessfully.
|
|
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
|
target_fs: target sample rate for synthesized audio.
|
|
|
|
|
wav_base64: The base64 format of the synthesized audio.
|
|
|
|
|
"""
|
|
|
|
|
|
|
|
|
|
# transform sample_rate
|
|
|
|
|
if target_fs == 0 or target_fs > original_fs:
|
|
|
|
|
target_fs = original_fs
|
|
|
|
|
wav_tar_fs = wav
|
|
|
|
|
logger.info(
|
|
|
|
|
"The sample rate of synthesized audio is the same as model, which is {}Hz".
|
|
|
|
|
format(original_fs))
|
|
|
|
|
else:
|
|
|
|
|
wav_tar_fs = librosa.resample(
|
|
|
|
|
np.squeeze(wav), original_fs, target_fs)
|
|
|
|
|
logger.info(
|
|
|
|
|
"The sample rate of model is {}Hz and the target sample rate is {}Hz. Converting the sample rate of the synthesized audio successfully.".
|
|
|
|
|
format(original_fs, target_fs))
|
|
|
|
|
# transform volume
|
|
|
|
|
wav_vol = wav_tar_fs * volume
|
|
|
|
|
logger.info("Transform the volume of the audio successfully.")
|
|
|
|
|
|
|
|
|
|
# transform speed
|
|
|
|
|
try: # windows not support soxbindings
|
|
|
|
|
wav_speed = change_speed(wav_vol, speed, target_fs)
|
|
|
|
|
logger.info("Transform the speed of the audio successfully.")
|
|
|
|
|
except ServerBaseException:
|
|
|
|
|
raise ServerBaseException(
|
|
|
|
|
ErrorCode.SERVER_INTERNAL_ERR,
|
|
|
|
|
"Failed to transform speed. Can not install soxbindings on your system. \
|
|
|
|
|
You need to set speed value 1.0.")
|
|
|
|
|
except BaseException:
|
|
|
|
|
logger.error("Failed to transform speed.")
|
|
|
|
|
|
|
|
|
|
# wav to base64
|
|
|
|
|
buf = io.BytesIO()
|
|
|
|
|
wavfile.write(buf, target_fs, wav_speed)
|
|
|
|
|
base64_bytes = base64.b64encode(buf.read())
|
|
|
|
|
wav_base64 = base64_bytes.decode('utf-8')
|
|
|
|
|
logger.info("Audio to string successfully.")
|
|
|
|
|
|
|
|
|
|
# save audio
|
|
|
|
|
if audio_path is not None:
|
|
|
|
|
if audio_path.endswith(".wav"):
|
|
|
|
|
sf.write(audio_path, wav_speed, target_fs)
|
|
|
|
|
elif audio_path.endswith(".pcm"):
|
|
|
|
|
wav_norm = wav_speed * (32767 / max(0.001,
|
|
|
|
|
np.max(np.abs(wav_speed))))
|
|
|
|
|
with open(audio_path, "wb") as f:
|
|
|
|
|
f.write(wav_norm.astype(np.int16))
|
|
|
|
|
logger.info("Save audio to {} successfully.".format(audio_path))
|
|
|
|
|
else:
|
|
|
|
|
logger.info("There is no need to save audio.")
|
|
|
|
|
|
|
|
|
|
return target_fs, wav_base64
|
|
|
|
|
|
|
|
|
|
def run(self,
|
|
|
|
|
sentence: str,
|
|
|
|
|
spk_id: int=0,
|
|
|
|
@ -275,31 +191,30 @@ class TTSEngine(BaseEngine):
|
|
|
|
|
save_path (str, optional): The save path of the synthesized audio.
|
|
|
|
|
None means do not save audio. Defaults to None.
|
|
|
|
|
|
|
|
|
|
Raises:
|
|
|
|
|
ServerBaseException: Throws an exception if tts inference unsuccessfully.
|
|
|
|
|
ServerBaseException: Throws an exception if postprocess unsuccessfully.
|
|
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
|
lang: model language
|
|
|
|
|
target_sample_rate: target sample rate for synthesized audio.
|
|
|
|
|
wav_base64: The base64 format of the synthesized audio.
|
|
|
|
|
"""
|
|
|
|
|
|
|
|
|
|
lang = self.config.lang
|
|
|
|
|
wav_list = []
|
|
|
|
|
|
|
|
|
|
for wav in self.executor.infer(text=sentence, lang=lang, am=self.config.am, spk_id=spk_id, am_block=self.am_block, am_pad=self.am_pad, voc_block=self.voc_block, voc_pad=self.voc_pad):
|
|
|
|
|
for wav in self.executor.infer(
|
|
|
|
|
text=sentence,
|
|
|
|
|
lang=lang,
|
|
|
|
|
am=self.config.am,
|
|
|
|
|
spk_id=spk_id,
|
|
|
|
|
am_block=self.am_block,
|
|
|
|
|
am_pad=self.am_pad,
|
|
|
|
|
voc_block=self.voc_block,
|
|
|
|
|
voc_pad=self.voc_pad):
|
|
|
|
|
# wav type: <class 'numpy.ndarray'> float32, convert to pcm (base64)
|
|
|
|
|
wav = float2pcm(wav) # float32 to int16
|
|
|
|
|
wav = float2pcm(wav) # float32 to int16
|
|
|
|
|
wav_bytes = wav.tobytes() # to bytes
|
|
|
|
|
wav_base64 = base64.b64encode(wav_bytes).decode('utf8') # to base64
|
|
|
|
|
wav_list.append(wav)
|
|
|
|
|
|
|
|
|
|
yield wav_base64
|
|
|
|
|
|
|
|
|
|
wav_all = np.concatenate(wav_list, axis=0)
|
|
|
|
|
logger.info("The durations of audio is: {} s".format(len(wav_all)/self.executor.am_config.fs))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
yield wav_base64
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
wav_all = np.concatenate(wav_list, axis=0)
|
|
|
|
|
logger.info("The durations of audio is: {} s".format(
|
|
|
|
|
len(wav_all) / self.executor.am_config.fs))
|
|
|
|
|