ONNX Runtime is a high performance inferencing and training engine for machine learning models. This show focuses on ONNX Runtime for model inference. ONNX Runtime has been widely adopted by a varietyof Microsoftproducts including Bing, Office365andAzureCognitiveServices,achieving an average of 2.9x inference speedup. Now we are glad to introduce ONNX Runtime quantization and ONNX Runtime mobile for further accelerating model inference with even smaller model size and runtime size. ONNX Runtime...

Read the full article at Microsoft MSDN Channel 9