fix ci and import

pull/538/head
Hui Zhang 5 years ago
parent 4da2852982
commit 84e0ea2342

@ -14,7 +14,7 @@
"""Test decoders.""" """Test decoders."""
import unittest import unittest
from decoders import decoders_deprecated as decoder from deepspeech.decoders import decoders_deprecated as decoder
class TestDecoders(unittest.TestCase): class TestDecoders(unittest.TestCase):

@ -13,7 +13,7 @@
# limitations under the License. # limitations under the License.
from yacs.config import CfgNode as CN from yacs.config import CfgNode as CN
from deepspeech.models.DeepSpeech2 import DeepSpeech2Model from deepspeech.models.deepspeech2 import DeepSpeech2Model
_C = CN() _C = CN()
_C.data = CN( _C.data = CN(

@ -44,7 +44,7 @@ from deepspeech.io.sampler import SortagradBatchSampler
from deepspeech.io.dataset import ManifestDataset from deepspeech.io.dataset import ManifestDataset
from deepspeech.training.loss import CTCLoss from deepspeech.training.loss import CTCLoss
from deepspeech.models.DeepSpeech2 import DeepSpeech2Model from deepspeech.models.deepspeech2 import DeepSpeech2Model
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)

@ -25,7 +25,7 @@ from paddle.nn import functional as F
from paddle.nn import initializer as I from paddle.nn import initializer as I
from deepspeech.modules.conv import ConvStack from deepspeech.modules.conv import ConvStack
from deepspeech.modules.conv import RNNStack from deepspeech.modules.rnn import RNNStack
from deepspeech.modules.mask import sequence_mask from deepspeech.modules.mask import sequence_mask
from deepspeech.modules.activation import brelu from deepspeech.modules.activation import brelu
from deepspeech.utils import checkpoint from deepspeech.utils import checkpoint
@ -79,7 +79,7 @@ class DeepSpeech2Model(nn.Layer):
share_rnn_weights=True #Whether to share input-hidden weights between forward and backward directional RNNs.Notice that for GRU, weight sharing is not supported. share_rnn_weights=True #Whether to share input-hidden weights between forward and backward directional RNNs.Notice that for GRU, weight sharing is not supported.
)) ))
if config is not None: if config is not None:
config.model.merge_from_other_cfg(default) config.merge_from_other_cfg(default)
return default return default
def __init__(self, def __init__(self,

@ -12,6 +12,7 @@
# See the License for the specific language governing permissions and # See the License for the specific language governing permissions and
# limitations under the License. # limitations under the License.
import math
import logging import logging
import paddle import paddle

@ -14,6 +14,7 @@
import logging import logging
import paddle
from paddle.fluid.dygraph import base as imperative_base from paddle.fluid.dygraph import base as imperative_base
from paddle.fluid import layers from paddle.fluid import layers
from paddle.fluid import core from paddle.fluid import core

Loading…
Cancel
Save