|
|
@ -591,7 +591,8 @@ def get_predictor(
|
|
|
|
|
|
|
|
|
|
|
|
config = inference.Config(
|
|
|
|
config = inference.Config(
|
|
|
|
str(Path(model_dir) / model_file), str(Path(model_dir) / params_file))
|
|
|
|
str(Path(model_dir) / model_file), str(Path(model_dir) / params_file))
|
|
|
|
config.enable_memory_optim()
|
|
|
|
if device != "npu":
|
|
|
|
|
|
|
|
config.enable_memory_optim()
|
|
|
|
config.switch_ir_optim(True)
|
|
|
|
config.switch_ir_optim(True)
|
|
|
|
if device == "gpu":
|
|
|
|
if device == "gpu":
|
|
|
|
config.enable_use_gpu(100, device_id)
|
|
|
|
config.enable_use_gpu(100, device_id)
|
|
|
|