ONNX Runtime Web unleashes generative AI in the browser using WebGPU
This blog is thrilled to announce the official launch of ONNX Runtime Web featuring WebGPU, now available in the ONNX Runtime 1.17 release. Read more
This blog is thrilled to announce the official launch of ONNX Runtime Web featuring WebGPU, now available in the ONNX Runtime 1.17 release. Read more
Continuing the ONNXRuntime On-Device Training blog series, we are introducing ONNX Runtime Training for Web, a new feature in ONNX Runtime (ORT) that enables training models in the browser. Read more
In this Blog we will describe our implementation of the DragGAN2 algorithm, based on StyleGAN1, using ONNX Runtime. We will give a technical overview of the architectures, describe the motivation and discuss challenges and their resolution. We have also released the python code that walks through the resulting implementation. Read more
Whether you’re a developer, an SRE, or trying to figure out how cloud-native technology can help deliver more for your business, wherever you go at KubeCon, you’ll find Microsoft building and serving the open-source communities that serve our customers. Read more
The Microsoft Azure Incubations team is excited to announce a new open application platform called Radius that places the application at the center of every workflow; redefining how applications are built, managed, and understood. Read more
LF AI & Data Foundation announced Recommenders as its latest Sandbox project. Read more
ONNX models can be accelerated with ONNX Runtime (ORT), which works cross-platform and provides coverage for many cloud models and language models. Updating the HF Model Hub with more accurate information about ONNX coverage will ensure that users can leverage all the benefits of ORT when deploying HF models. Read more
ONNX Runtime, Intel®, and Microsoft developed the 8-bit integer matrix multiplication kernel in ONNX Runtime using Intel® AMX instructions, resulting in four times faster performance than 3rd Gen Intel® Xeon® using Intel® DL Boost. Read more
KEDA reduces the complexity of infrastructure autoscaling, it makes it simpler for Kubernetes cluster administrators to configure, manage, and secure the application auto-scaler. Read more
ONNX Script is a new open-source library for directly authoring ONNX models in Python with a focus on clean, idiomatic Python syntax and composability through ONNX-native functions. Read more
By thinking outside the box, we can envision creating a virtual multiverse. Within this innovative space, one can propose, evaluate, and decide on multiple hypotheses. Real-world examples of this approach include planning new product configurations, operating a plant, designing heating or cooling systems, or responding to catastrophes. Read more
Building upon the foundation we established earlier, this blog will present comprehensive information about the underlying details of training models directly on user devices using ORT. Equipped with these technical details, we encourage you to try out On-Device Training with ONNX Runtime for your custom scenario. Read more