ML Inference on Edge devices with ONNX Runtime using Azure DevOps
AI applications are designed to perform tasks that emulate human intelligence to make predictions that help us make better decisions for the scenario. This drives operational efficiency when the machine executes the task without worrying about fatigue or safety. But the effectiveness of the AI application... Read more
Accelerate your NLP pipelines using Hugging Face Transformers and ONNX Runtime
This post was written by Morgan Funtowicz from Hugging Face and Tianlei Wu from Microsoft Transformer models have taken the world of natural language processing (NLP) by storm. They went from beating all the research benchmarks to getting adopted for production by a growing number of companies in a record... Read more
There are tons of options in the Machine Learning world. You may have noticed a large number of frameworks, libraries, and formats that are floating around. Machine Learning is one of the most “hot” topics of the decade, and the tooling built as we are riding... Read more
Accelerating Model Training with the ONNX Runtime
TDLR; This article introduces the new improvements to the ONNX runtime for accelerated training and outlines the 4 key steps for speeding up training of an existing PyTorch model with the ONNX Runtime (ORT). What is the ONNX Runtime (ORT)? ONNX Runtime is a performance-focused inference engine... Read more
Tutorial: Accelerate and Productionize ML Model Inferencing Using Open-Source Tools
Faith and Prabhat are speakers for ODSC East 2020 this April. Be sure to check out their talk, “From Research to Production: Performant Cross-platform ML/DNN Model Inferencing on Cloud and Edge with ONNX Runtime,” there! You’ve finally got that perfect trained model for your data set.... Read more