There’s a huge wealth of open-source software components available today for embedding vision on the latest SoCs from suppliers such as NXP, Broadcom, TI and NVIDIA, at lower power and cost points than ever before. Testing vision algorithms is the first step, but what about the rest of your system? In this talk, we will consider the best open-source components available today and explain how to select and integrate them to build complete video pipelines on Linux – from camera to display – while maximizing performance. We’ll examine and compare popular open-source libraries for vision, including Yocto, ffmpeg, gstreamer, V4L2, OpenCV, OpenVX, OpenCL and OpenGL. Which components do you need and why? We will also summarize the steps required to build and test complete video pipelines, common integration problems to avoid and how to work around issues to get the best performance possible on embedded systems.