Intel Deep Learning Deployment Toolkit 【PREMIUM】
The easiest way to get the runtime is via pip, though for the full Model Optimizer, download the full OpenVINO toolkit.
Take your slowest production model, run it through the Model Optimizer, and benchmark the result. You will be shocked. Have you used OpenVINO or the Intel DLDT in production? Let me know your latency improvements in the comments below! intel deep learning deployment toolkit
If you are deploying to CPUs (and let's be honest, 90% of inference still happens on CPUs), you are leaving performance on the table by not using DLDT. The easiest way to get the runtime is
The toolkit solves one simple problem:
Stop wrestling with framework dependencies. Start deploying optimized models at the edge. If you have ever trained a beautiful model in PyTorch or TensorFlow only to watch it crawl across the finish line on a production CPU, you know the pain. We’ve all been there: high latency, bloated memory usage, and the sinking feeling that you need to buy expensive GPUs just to serve inference. Have you used OpenVINO or the Intel DLDT in production