AI is on the cusp of a revolution, driven by the convergence of several breakthroughs. These include large language models that can reason like humans, and the integration of natural language and computer vision through vision-language models (VLMs). In this keynote talk, Professor Trevor Darrell will present his group’s groundbreaking work, including methods for training vision models when labeled data is unavailable and techniques that enable robots to determine appropriate actions in novel situations. Much of Darrell’s work aims to overcome obstacles—such as massive memory and compute requirements—that limit the practical applications of state-of-the-art models. He will discuss approaches to making VLMs smaller and more efficient while retaining accuracy. He will also show how LLMs can be used as visual reasoning coordinators, overseeing the use of multiple task-specific models to enable superior performance.
Gérard Medioni, Vice President and Distinguished Scientist at Prime Video and Amazon MGM Studios, will discuss his work on the innovative Just Walk Out technology, as well as the Amazon One identity service. Moving to the world of entertainment, the session will also highlight the technology that powers Prime Video, including AI innovations that are improving the streaming experience for over 200 million Prime members worldwide. Attendees will gain insights into how these technologies are reshaping entertainment, and consider how they will evolve in the coming years to enhance viewer engagement, storytelling and personalization.
Learn about the newest speakers, sessions and other noteworthy details about the Summit Program by leaving us a few details.
Interested in sponsoring or exhibiting?
The Embedded Vision Summit gives you unique access to the best qualified technology buyers you’ll ever meet.
Want to contact us?
Use the small blue chat widget in the lower right-hand corner of your screen, or the form linked below.
STAY CONNECTED
Follow us on Twitter and LinkedIn.