There are few things that can increase an organization’s bottom line like an Edge AI implementation that delivers on its potential. Imagine a tractor enabled with AI computer vision that needs to limit its speed to 20mph to be able to recognize what is a weed and what is a crop. If the inference speed for weed detection can be increased by only 10%, the actual real-world gains mean the tractor can double its speed to 40 mph while also doubling its production and reducing overall fuel use.
Edge AI is so promising because it can enable organizations to make exponential gains from incremental improvements. The problem is that most organizations are so lured by the promise of AI that they rush past how littered the landscape is with projects that have failed. According to Gartner, only 53% of projects make it to production. In fact, a full 85% of AI projects fail to meet their initial business goals. Those numbers reveal exactly how challenging Edge AI implementations are. Because Edge AI requires highly optimized models to work efficiently, ML Engineers have to go through a frustrating process that includes:
- Overcoming resource constraints and making trade-offs between model accuracy, inference speed, size, and memory. Power and cost concerns only add to optimization challenges.
- Tackling the idiosyncrasies of different hardware, compilers and development frameworks. In fact, two devices from the same family can routinely return entirely different results when using the same optimization algorithms on the same model.
- Diagnosing and resolving issues caused by a continuously changing data and model landscape as real-world feedback is received.
The easier way to Edge AI
We knew organizations needed a way to improve their odds of Edge AI success. We realized that by automating the ML optimization pipeline that we could resolve Edge AI deployment challenges and let organizations instead focus on their business outcomes. We created the Latent AI Efficient Inference Platform (LEIP) to empower AI developers with adaptive, on-premise software tools that can compile, compress, and deploy AI models for any hardware target, framework, or OS.
While LEIP provides ML engineers with everything they need for success, the process still requires configuration to get the best results. Quantization algorithms (symmetric vs asymmetric vs tensor-based vs channel-based, etc.), data layouts (channel-first-NCHW vs channel-last-NHWC), and compiler optimizations (graph partitioning, operator fusion, math kernel selection) all have to be configured while additional techniques like Tensor Splitting and Bias Correction often have to be added.
We knew we wanted to go further to help our customers succeed. We created Latent AI LEIP Recipes to take the stress out of Edge AI deployments. Latent AI LEIP Recipes are a set of pre-configured assets combined with a set of instructions to follow to get to an optimized model. Each LEIP Recipe tackles a type of problem like object detection or classification and is configured for a specific model and hardware target. It comes with all the model optimization settings pre-configured including quantization, structured pruning and throttling, and lets ML engineers leverage models pre-optimized for low-power, inference speed, size and memory.
Accelerate your time to market
Because our goal is to ensure your Edge AI projects can drive timely business decisions, one of our key performance metrics is your time to market. A typical ML Engineer spends considerable time resolving the ML model into a format that can run on the hardware. Without accounting for training time, it can take up to 3 days to deploy a standard object detection model like the YoloV5 large when optimized using Nvidia TRT libraries to an AGX. However, a LEIP Recipe can accelerate your time to market by 10x by reducing this process to a few hours.
With LEIP Recipes, ML engineers focus on optimization using pre-configured settings on the LEIP SDK. LEIP Recipes have several other advantages including:
- Highly optimized models – When we trained a recipe using the widely popular YoloV5 large model on the COCO dataset and optimized it for the Nvidia AGX, the results were excellent compared to unoptimized models. We showed more than a 3x improvement in inference speed when compared to a Pytorch baseline for a Nvidia Xavier AGX processor. Simultaneously, we observed an 84% reduction in model size while only seeing a slight accuracy drop below 3%.
- Exceptional runtime support – LEIP Recipes support both Python and C++ runtimes out of the box to make sure integrating with your hardware is easy and convenient. You can also manage any changes to inputs and outputs of your model by easily swapping out your pre-processing and post-processing functions in your LEIP Recipe.
- Robust, Repeatable, Reproducible models – Because LEIP Recipes are built using software engineering fundamentals like modularity and composability, they can be configured with YAML files. And they run via a docker container that is packaged with all its dependencies included so you can run the Recipes anywhere, anytime.
- Scaling innovation – For larger ML teams and System Integrators, LEIP Recipes are a mechanism to scale quickly. Over 50+ combinations of models and hardware targets are immediately available to train and deploy with your data.
For organizations who need a better way to implement AI than the cloud can offer, Latent AI Recipes enable Edge AI solutions with reduced power, latency, and memory footprints while maintaining model accuracy. Latent AI LEIP Recipes can help you overcome the steep challenges Edge AI presents and bake in a plan for success. To learn more about how LEIP Recipes can help you gain actionable, reliable, and repeatable insights from your data, or to inquire about a specific configuration, contact us at [email protected].