- Execution Provider updates
- MKL-DNN provider (subgraph based execution) for improved performance
- Intel OpenVINO EP now available for Public Preview - build instructions
- Update to CUDA 10 for inferencing with NVIDIA GPUs
- Base CPU EP has faster convolution performance using the NCHWc blocked layout. This layout optimization can be enabled by setting graph optimization level to 3 in the session options.
- C++ API for inferencing (wrapper on C API)
- ONNX Runtime Server (Beta) for inferencing with HTTP and GRPC endpoints
- Python Operator (Beta) to support custom Python code in a single node of an ONNX graph to make it easier for experimentation of custom operators
- Support of Keras-based Mask R-CNN model. The model relies on some custom operators pending to be added in ONNX; in the meantime, it can be converted using this script for inferencing using ONNX Runtime 0.5. Other object detection models can be found from the ONNX Model Zoo.
- Minor updates to the C API
- For consistency, all C APIs now return an ORT status code
- Code coverage for this release is 83%