Date: Tuesday, May 12
Start Time: 10:25 am
End Time: 11:10 am
Vision-language models are moving fast, but it’s not always clear where they add value, and many teams struggle to turn demos into dependable products. In this plenary panel, we’ll cut through the hype and focus on where VLMs make sense and the barriers to deploying them in real systems. We’ll discuss where VLMs are delivering clear value and where classic CV still wins on cost, latency and determinism. Panelists will compare practical hybrid architectures, examine failure modes such as weak grounding and hallucination and outline guardrails, evaluation methods and monitoring that work in production. We’ll also debate the edge vs. cloud split, domain adaptation strategies and the privacy/security governance required when models can answer open-ended questions about people and places. Attendees will leave with insights into requirements and potential pitfalls to de-risk their VLM road map.





