Category "onnxruntime"

Onnxruntime NodeJS set intraOpNumThreads and interOpNumThreads by execution mode

I'm using Onnxruntime in NodeJS to execute onnx converted models in cpu backend to run inference. According to the docs, the optional parameters are the followi

failed to inference ONNX model: TypeError: Cannot read properties of undefined (reading 'InferenceSession')

I tried to replicate the example found here: https://github.com/microsoft/onnxruntime-inference-examples/tree/main/js/quick-start_onnxruntime-web-bundler: impor

Can't create an inference session in onnxruntime

I am not able to create an instance of InferenceSession using onnxruntime. My platform is Mac OS(Big Sur). The code doesn't even throw any exceptions. Process i

Building models in ONNX

Is it possible to build a model in ONNX without using a different deep learning framework (e.g. PyTorch, TensorFlow, etc.)? In PyTorch, I would write a model l

OnnxRuntime vs OnnxRuntime+OpenVinoEP inference time difference

I'm trying to accelerate my model's performance by converting it to OnnxRuntime. However, I'm getting weird results, when trying to measure inference time. Whil