From af65d5aa0c9d3be8f742f5693ac61c5be3c805f0 Mon Sep 17 00:00:00 2001 From: huangyuxin Date: Sun, 17 Oct 2021 02:23:52 +0000 Subject: [PATCH] revert the ctc_loss in deepspeech/__init__.py --- deepspeech/__init__.py | 14 +++----------- 1 file changed, 3 insertions(+), 11 deletions(-) diff --git a/deepspeech/__init__.py b/deepspeech/__init__.py index 90ab2223..493f10a6 100644 --- a/deepspeech/__init__.py +++ b/deepspeech/__init__.py @@ -362,19 +362,11 @@ def ctc_loss(logits, label_lengths, blank=0, reduction='mean', - norm_by_times=True, - norm_by_batchsize=False, - norm_by_total_logits_len=False): + norm_by_times=True): #logger.info("my ctc loss with norm by times") ## https://github.com/PaddlePaddle/Paddle/blob/f5ca2db2cc/paddle/fluid/operators/warpctc_op.h#L403 - loss_out = paddle.fluid.layers.warpctc( - logits, - labels, - blank, - norm_by_times, - input_lengths, - label_lengths, - norm_by_batchsize, ) + loss_out = paddle.fluid.layers.warpctc(logits, labels, blank, norm_by_times, + input_lengths, label_lengths) loss_out = paddle.fluid.layers.squeeze(loss_out, [-1]) assert reduction in ['mean', 'sum', 'none']