|
import onnxruntime |
|
import torch |
|
|
|
providers = [ |
|
|
|
|
|
|
|
|
|
|
|
|
|
('CUDAExecutionProvider', { |
|
'device_id': 0, |
|
'arena_extend_strategy': 'kSameAsRequested', |
|
'gpu_mem_limit': 8 * 1024 * 1024 * 1024, |
|
'cudnn_conv_algo_search': 'HEURISTIC', |
|
}) |
|
] |
|
|
|
def load_onnx(file_path: str): |
|
assert file_path.endswith(".onnx") |
|
sess_opt = onnxruntime.SessionOptions() |
|
ort_session = onnxruntime.InferenceSession(file_path, sess_opt=sess_opt, providers=providers) |
|
return ort_session |
|
|
|
|
|
def load_onnx_caller(file_path: str, single_output=False): |
|
ort_session = load_onnx(file_path) |
|
def caller(*args): |
|
torch_input = isinstance(args[0], torch.Tensor) |
|
if torch_input: |
|
torch_input_dtype = args[0].dtype |
|
torch_input_device = args[0].device |
|
|
|
assert all([isinstance(arg, torch.Tensor) for arg in args]), "All inputs should be torch.Tensor, if first input is torch.Tensor" |
|
assert all([arg.dtype == torch_input_dtype for arg in args]), "All inputs should have same dtype, if first input is torch.Tensor" |
|
assert all([arg.device == torch_input_device for arg in args]), "All inputs should have same device, if first input is torch.Tensor" |
|
args = [arg.cpu().float().numpy() for arg in args] |
|
|
|
ort_inputs = {ort_session.get_inputs()[idx].name: args[idx] for idx in range(len(args))} |
|
ort_outs = ort_session.run(None, ort_inputs) |
|
|
|
if torch_input: |
|
ort_outs = [torch.tensor(ort_out, dtype=torch_input_dtype, device=torch_input_device) for ort_out in ort_outs] |
|
|
|
if single_output: |
|
return ort_outs[0] |
|
return ort_outs |
|
return caller |
|
|