Intel Neural Compressor
Shares tags: deploy, hardware & accelerators, cpu-only optimizers
Effortless model compression and deployment for CPUs, GPUs, and NPUs.
Tags
Similar Tools
Other tools you might consider
Intel Neural Compressor
Shares tags: deploy, hardware & accelerators, cpu-only optimizers
ONNX Runtime CPU EP
Shares tags: deploy, hardware & accelerators, cpu-only optimizers
Intel OpenVINO
Shares tags: deploy, hardware & accelerators, cpu-only optimizers
Apache TVM Unity
Shares tags: deploy, hardware & accelerators, cpu-only optimizers
overview
Microsoft Olive is a powerful model compression and compilation pipeline designed to enhance the efficiency of AI models. It targets a variety of hardware platforms, including CPUs, GPUs, and NPUs, helping developers deploy models with ease.
features
Olive offers a plethora of features aimed at improving model performance while simplifying the optimization process. With new updates in mixed-precision and quantization, users can expect enhanced efficiency and reduced operational costs.
use_cases
Olive is invaluable for AI model developers and machine learning engineers seeking to enhance model deployment across various hardware platforms. Whether you're working in the cloud or on edge devices, Olive simplifies the optimization lifecycle.
Microsoft Olive is designed for AI model developers, machine learning engineers, and organizations that want to optimize deployment of ONNX models across various hardware platforms.
Recent updates include selective mixed precision, blockwise RTN quantization, and a native GPTQ implementation, all geared towards improving model efficiency and lowering deployment costs.
With enhancements such as a streamlined CLI command, improved Python APIs, and advanced tools for evaluation and packaging, Olive enables users to optimize and test models with minimal manual intervention.