diff --git a/paddlespeech/cls/exps/panns/deploy/predict.py b/paddlespeech/cls/exps/panns/deploy/predict.py index fcc2822d0..2888590c9 100644 --- a/paddlespeech/cls/exps/panns/deploy/predict.py +++ b/paddlespeech/cls/exps/panns/deploy/predict.py @@ -56,7 +56,7 @@ def extract_features(files: str, **kwargs): feature_extractor = LogMelSpectrogram(sr, **kwargs) feat = feature_extractor(paddle.to_tensor(waveforms[i])) - feat = paddle.transpose(feat, perm=[1, 0]).unsqueeze(0) + feat = paddle.transpose(feat, perm=[1, 0]) feats.append(feat) @@ -86,6 +86,7 @@ class Predictor(object): params_file), 'Please check model and parameter files.' config = inference.Config(model_file, params_file) + config.disable_mkldnn() if device == "gpu": # set GPU configs accordingly # such as intialize the gpu memory, enable tensorrt diff --git a/paddlespeech/cls/exps/panns/export_model.py b/paddlespeech/cls/exps/panns/export_model.py index 63b22981a..e860b54aa 100644 --- a/paddlespeech/cls/exps/panns/export_model.py +++ b/paddlespeech/cls/exps/panns/export_model.py @@ -39,7 +39,8 @@ if __name__ == '__main__': input_spec=[ paddle.static.InputSpec( shape=[None, None, 64], dtype=paddle.float32) - ]) + ], + full_graph=True) # Save in static graph model. paddle.jit.save(model, os.path.join(args.output_dir, "inference"))