Improve BERT inference speed by combining the power of Optimum, OpenVINO™, ONNX Runtime, and Azure 

5 min read

Make large models smaller and faster with OpenVino Execution Provider, NNCF and ONNX Runtime leveraging Azure Machine Learning. Read more

Faster inference for PyTorch models with OpenVINO Integration with Torch-ORT 

4 min read

Many developers opt to use popular AI Frameworks like PyTorch, which simplifies the process of analyzing predictions, training models, leveraging data, and refining future results. Read more