OpenVINO

Introduction Deep learning tools are extensively used in vision applications across most industries, ranging from facial recognition on mobile devices to Tesla’s self-driving cars. However, using the right tools is

Guest post by Aleksandr Voron To keep OpenVINO™ Toolkit focused on optimizing and deploying inference, we no longer include OpenCV and DL Streamer in our distribution packages. But not to

Deep Learning Inference Engine backend from the Intel OpenVINO toolkit is one of the supported OpenCV DNN backends. It was mentioned in the previous post that ARM CPUs support has

We are pleased to announce that Deep Learning Inference Engine backend in the OpenCV DNN module can run inference on ARM CPUs nowadays. Previously Inference Engine supported Intel hardware only:

We have already discussed several ways to convert your DL model into OpenVINO in previous blogs (PyTorch and TensorFlow). Let’s try something more advanced now.
How TensorFlow trained model may be used and deployed to run with OpenVINO Inference Engine

Are you looking for a fast way to run neural network inferences on Intel platforms? Then OpenVINO toolkit is exactly what you need. It provides a large number of optimizations

Nowadays, many ground-breaking solutions based on neural network are developed daily and more people are adopting this technique for solving problems such as voice recognitions in their life. Because of

How TensorFlow trained model may be used and deployed to run with OpenVINO Inference Engine

Join the waitlist to receive a 20% discount

Courses are (a little) oversubscribed and we apologize for your enrollment delay. As an apology, you will receive a 20% discount on all waitlist course purchases. Current wait time will be sent to you in the confirmation email. Thank you!