Improve BERT inference speed by combining the power of Optimum, OpenVINO™, ONNX Runtime, and Azure 

5 min read

Make large models smaller and faster with OpenVino Execution Provider, NNCF and ONNX Runtime leveraging Azure Machine Learning. Read more