The adoption of large language models (LLMs) and vision-language models (VLMs) in edge AI applications has been limited by on-device computation and memory resources, and also by the cost of training and developing large models suitable for edge applications. The adoption of new training methods and algorithms combined with innovative new model architectures promises to dramatically change the outlook for adoption with models created for the edge that can efficiently execute within SWaP and cost targets.