change CodeStyle, test=asr

pull/2755/head
yeyupiaoling 3 years ago
parent 6d867f714d
commit 2aa84571c0

@ -42,7 +42,8 @@ from paddlespeech.s2t.modules.cmvn import GlobalCMVN
from paddlespeech.s2t.modules.ctc import CTCDecoderBase from paddlespeech.s2t.modules.ctc import CTCDecoderBase
from paddlespeech.s2t.modules.decoder import BiTransformerDecoder from paddlespeech.s2t.modules.decoder import BiTransformerDecoder
from paddlespeech.s2t.modules.decoder import TransformerDecoder from paddlespeech.s2t.modules.decoder import TransformerDecoder
from paddlespeech.s2t.modules.encoder import ConformerEncoder, SqueezeformerEncoder from paddlespeech.s2t.modules.encoder import ConformerEncoder
from paddlespeech.s2t.modules.encoder import SqueezeformerEncoder
from paddlespeech.s2t.modules.encoder import TransformerEncoder from paddlespeech.s2t.modules.encoder import TransformerEncoder
from paddlespeech.s2t.modules.initializer import DefaultInitializerContext from paddlespeech.s2t.modules.initializer import DefaultInitializerContext
from paddlespeech.s2t.modules.loss import LabelSmoothingLoss from paddlespeech.s2t.modules.loss import LabelSmoothingLoss

@ -26,7 +26,7 @@ from paddlespeech.s2t.utils.log import Log
logger = Log(__name__).getlog() logger = Log(__name__).getlog()
__all__ = ["MultiHeadedAttention", "RelPositionMultiHeadedAttention"] __all__ = ["MultiHeadedAttention", "RelPositionMultiHeadedAttention", "RelPositionMultiHeadedAttention2"]
# Relative Positional Encodings # Relative Positional Encodings
# https://www.jianshu.com/p/c0608efcc26f # https://www.jianshu.com/p/c0608efcc26f

@ -5,7 +5,7 @@ from paddle import nn
from paddle.nn import initializer as I from paddle.nn import initializer as I
from typeguard import check_argument_types from typeguard import check_argument_types
__all__ = ['ConvolutionModule'] __all__ = ['ConvolutionModule2']
from paddlespeech.s2t import masked_fill from paddlespeech.s2t import masked_fill
from paddlespeech.s2t.modules.align import Conv1D, BatchNorm1D, LayerNorm from paddlespeech.s2t.modules.align import Conv1D, BatchNorm1D, LayerNorm

@ -43,7 +43,7 @@ from paddlespeech.s2t.utils.log import Log
logger = Log(__name__).getlog() logger = Log(__name__).getlog()
__all__ = ["BaseEncoder", 'TransformerEncoder', "ConformerEncoder"] __all__ = ["BaseEncoder", 'TransformerEncoder', "ConformerEncoder", "SqueezeformerEncoder"]
class BaseEncoder(nn.Layer): class BaseEncoder(nn.Layer):

@ -26,7 +26,7 @@ from paddlespeech.s2t.utils.log import Log
logger = Log(__name__).getlog() logger = Log(__name__).getlog()
__all__ = ["TransformerEncoderLayer", "ConformerEncoderLayer"] __all__ = ["TransformerEncoderLayer", "ConformerEncoderLayer", "SqueezeformerEncoderLayer"]
class TransformerEncoderLayer(nn.Layer): class TransformerEncoderLayer(nn.Layer):

@ -23,7 +23,7 @@ from paddlespeech.s2t.utils.log import Log
logger = Log(__name__).getlog() logger = Log(__name__).getlog()
__all__ = ["PositionwiseFeedForward"] __all__ = ["PositionwiseFeedForward", "PositionwiseFeedForward2"]
class PositionwiseFeedForward(nn.Layer): class PositionwiseFeedForward(nn.Layer):

@ -32,7 +32,8 @@ logger = Log(__name__).getlog()
__all__ = [ __all__ = [
"LinearNoSubsampling", "Conv2dSubsampling4", "Conv2dSubsampling6", "LinearNoSubsampling", "Conv2dSubsampling4", "Conv2dSubsampling6",
"Conv2dSubsampling8" "Conv2dSubsampling8", "TimeReductionLayerStream", "TimeReductionLayer1D",
"TimeReductionLayer2D", "DepthwiseConv2DSubsampling4"
] ]

Loading…
Cancel
Save