@@ -28,6 +28,7 @@ def export(args, model=None, save_dir=None, use_ema=False):
28
28
assert args .config is not None , \
29
29
'No configuration file specified, please set --config'
30
30
cfg = Config (args .config )
31
+ use_fd_inference = True
31
32
if not model :
32
33
# save model
33
34
builder = SegBuilder (cfg )
@@ -38,6 +39,7 @@ def export(args, model=None, save_dir=None, use_ema=False):
38
39
logger .info ('Loaded trained params successfully.' )
39
40
if args .output_op != 'none' :
40
41
model = WrappedModel (model , args .output_op )
42
+ use_fd_inference = False
41
43
utils .show_env_info ()
42
44
utils .show_cfg_info (cfg )
43
45
else :
@@ -62,7 +64,7 @@ def export(args, model=None, save_dir=None, use_ema=False):
62
64
save_name = 'model'
63
65
yaml_name = 'deploy.yaml'
64
66
65
- if uniform_output_enabled :
67
+ if uniform_output_enabled and use_fd_inference == True :
66
68
inference_model_path = os .path .join (save_dir , "inference" , save_name )
67
69
yml_file = os .path .join (save_dir , "inference" , yaml_name )
68
70
if use_ema :
@@ -95,21 +97,27 @@ def export(args, model=None, save_dir=None, use_ema=False):
95
97
if cfg .dic .get ("pdx_model_name" , None ):
96
98
deploy_info ["Global" ] = {}
97
99
deploy_info ["Global" ]["model_name" ] = cfg .dic ["pdx_model_name" ]
98
- if cfg .dic .get ("hpi_config_path" , None ):
99
- with open (cfg .dic ["hpi_config_path" ], "r" ) as fp :
100
- hpi_config = yaml .load (fp , Loader = yaml .SafeLoader )
101
- if hpi_config ["Hpi" ]["backend_config" ].get ("paddle_tensorrt" , None ):
102
- hpi_config ["Hpi" ]["supported_backends" ]["gpu" ].remove (
103
- "paddle_tensorrt" )
104
- del hpi_config ['Hpi' ]['backend_config' ]['paddle_tensorrt' ]
105
- if hpi_config ["Hpi" ]["backend_config" ].get ("tensorrt" , None ):
106
- hpi_config ["Hpi" ]["supported_backends" ]["gpu" ].remove ("tensorrt" )
107
- del hpi_config ['Hpi' ]['backend_config' ]['tensorrt' ]
108
- hpi_config ["Hpi" ]["selected_backends" ]["gpu" ] = "paddle_infer"
109
- deploy_info ["Hpi" ] = hpi_config ["Hpi" ]
100
+ if cfg .dic .get ("uniform_output_enabled" , False ):
101
+ dynamic_shapes = {
102
+ 'x' : [[1 , 3 , 128 , 256 ], [1 , 3 , 512 , 1024 ], [1 , 3 , 1024 , 2048 ]]
103
+ }
104
+ supported_batch_size = [1 , 100 ]
105
+
106
+ backend_keys = ['paddle_infer' , 'tensorrt' ]
107
+ hpi_config = {
108
+ "backend_configs" : {
109
+ key : {
110
+ "dynamic_shapes" if key == "tensorrt" else "trt_dynamic_shapes" :
111
+ dynamic_shapes
112
+ }
113
+ for key in backend_keys
114
+ }
115
+ }
116
+ deploy_info ["Hpi" ] = hpi_config
110
117
msg = '\n ---------------Deploy Information---------------\n '
111
118
msg += str (yaml .dump (deploy_info ))
112
- logger .info (msg )
119
+ if use_fd_inference == False :
120
+ logger .info (msg )
113
121
114
122
with open (yml_file , 'w' ) as file :
115
123
yaml .dump (deploy_info , file )
0 commit comments