Multimodal LLMs promise to bring exciting new abilities to devices! As we see foundational models become more capable, we see compute requirements grow as well. It is not uncommon to see LLMs grow to tens of billions of parameters, at […]
Multimodal LLMs promise to bring exciting new abilities to devices! As we see foundational models become more capable, we see compute requirements grow as well. It is not uncommon to see LLMs grow to tens of billions of parameters, at […]
Many computer vision projects reach proof of concept but stall before production due to high costs, deployment challenges and infrastructure complexity. This session explores the path from prototype to production, focusing on how to reduce the cost of vision workloads […]
Large language models (LLMs) often demand hand-coded conversion scripts for deployment on each distinct processor-specific software stack—a process that’s time-consuming and prone to error. In this session, we introduce a model-agnostic approach designed to streamline LLM deployment, especially for the […]
At the embedded edge, choices of language model architectures have profound implications on the ability to meet demanding performance, latency and energy efficiency requirements. In this presentation, we contrast state-space models (SSMs) with transformers for use in this constrained regime. […]
Quadric’s Chimera general-purpose neural processor executes complete AI/ML graphs—all layers, including pre- and post-processing functions traditionally run on separate DSP processors. To enable this, the Chimera Graph Compiler processes and optimizes a combination of NN graphs, Python code and C++ […]
As humans, when we look at a scene our first impressions are sometimes wrong; we need to take a second look, to squint and reassess. Squinting enables us to focus our attention on the subject we are investigating and often […]
In this talk we’ll present Moonshine, a speech-to-text model that outperforms OpenAI’s Whisper by a factor of five in terms of speed. Leveraging this efficiency, we’ll show how to build a voice interface on a low-cost, resource-constrained Cortex-A SoC using […]
As large language models (LLMs) and vision-language models (VLMs) have quickly become important for edge applications from smartphones to automobiles, chipmakers and IP providers have struggled with how to adapt processor software stacks. In this talk, Expedera’s Ram Tadishetti examines […]
True innovation in tiny machine learning (tinyML) emerges from a synergy between software ingenuity, real-world application insights and leading-edge processor IP. In this presentation, we will explore the process of integrating these elements to shape the design of our latest […]
The deployment of neural networks near sensors brings well-known advantages such as lower latency, privacy and reduced overall system cost—but also brings significant challenges that complicate development. These challenges can be addressed effectively by choosing the right solution and design […]
Interested in sponsoring or exhibiting?
The Embedded Vision Summit gives you unique access to the best qualified technology buyers you’ll ever meet.
Want to contact us?
Use the small blue chat widget in the lower right-hand corner of your screen, or the form linked below.
STAY CONNECTED
Follow us on Twitter and LinkedIn.