in optimum/amd/ryzenai/modeling.py [0:0]
def _prepare_onnx_outputs(self, onnx_outputs, use_torch: bool):
outputs = {}
# converts onnxruntime outputs into tensor for standard outputs
for output, idx in self.output_names.items():
outputs[output] = onnx_outputs[idx]
if use_torch:
outputs[output] = torch.from_numpy(outputs[output])
return outputs