Overview

Overview

The Arm ML Inference Advisor (Arm MLIA) helps AI developers design and optimize neural network models for efficient inference on Arm targets. MLIA enables insights into how the ML model will perform on Arm early in the model development cycle. By passing a model file and specifying an Arm hardware target, users get an overview of possible areas of improvement and an actionable advice on how to address each of them. The advice can cover operator compatibility, performance analysis and model optimization (e.g. pruning and clustering). With the Arm ML Inference Advisor, we aim to make the Arm ML IP accessible to developers at all levels of abstraction, with differing knowledge on hardware optimization and machine learning.  The figure below gives an overview of the data flow.

MLIA data flow overview

For further technical information, refer to the documentation on the pypi.org homepage for Arm MLIA. For a quick introduction to the tool, check out this blog post.