Merge pull request #1697 from buchongyu2/develop

修改hack 单词拼写错误
pull/1698/head
Hui Zhang 3 years ago committed by GitHub
commit 8f1335d447
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23

@ -26,10 +26,10 @@ from paddlespeech.s2t.utils.log import Log
#TODO(Hui Zhang): remove fluid import #TODO(Hui Zhang): remove fluid import
logger = Log(__name__).getlog() logger = Log(__name__).getlog()
########### hcak logging ############# ########### hack logging #############
logger.warn = logger.warning logger.warn = logger.warning
########### hcak paddle ############# ########### hack paddle #############
paddle.half = 'float16' paddle.half = 'float16'
paddle.float = 'float32' paddle.float = 'float32'
paddle.double = 'float64' paddle.double = 'float64'
@ -110,7 +110,7 @@ if not hasattr(paddle, 'cat'):
paddle.cat = cat paddle.cat = cat
########### hcak paddle.Tensor ############# ########### hack paddle.Tensor #############
def item(x: paddle.Tensor): def item(x: paddle.Tensor):
return x.numpy().item() return x.numpy().item()
@ -353,7 +353,7 @@ if not hasattr(paddle.Tensor, 'tolist'):
setattr(paddle.Tensor, 'tolist', tolist) setattr(paddle.Tensor, 'tolist', tolist)
########### hcak paddle.nn ############# ########### hack paddle.nn #############
class GLU(nn.Layer): class GLU(nn.Layer):
"""Gated Linear Units (GLU) Layer""" """Gated Linear Units (GLU) Layer"""

@ -26,10 +26,10 @@ from paddlespeech.s2t.utils.log import Log
#TODO(Hui Zhang): remove fluid import #TODO(Hui Zhang): remove fluid import
logger = Log(__name__).getlog() logger = Log(__name__).getlog()
########### hcak logging ############# ########### hack logging #############
logger.warn = logger.warning logger.warn = logger.warning
########### hcak paddle ############# ########### hack paddle #############
paddle.half = 'float16' paddle.half = 'float16'
paddle.float = 'float32' paddle.float = 'float32'
paddle.double = 'float64' paddle.double = 'float64'
@ -110,7 +110,7 @@ if not hasattr(paddle, 'cat'):
paddle.cat = cat paddle.cat = cat
########### hcak paddle.Tensor ############# ########### hack paddle.Tensor #############
def item(x: paddle.Tensor): def item(x: paddle.Tensor):
return x.numpy().item() return x.numpy().item()

Loading…
Cancel
Save