LLMs On The Edge

What’s needed to utilize advances in AI in small devices.

popularity

Nearly all the data input for AI so far has been text, but that’s about to change. In the future, that input likely will include video, voice, as well as other types of data, causing a massive increase in the amount of data that needs to be modeled and the compute resources necessary to make it all work. This is hard enough in hyperscale data centers, which are sprouting up everywhere to handle the training and some inferencing, but it’s even more of a challenge in bandwidth- and power-limited edge devices. Sharad Chole, chief scientist and co-founder of Expedera, examines the tradeoffs involved in making this work, how to reduce the size of LLMs, and what impact this will have on engineers working in this space.



Leave a Reply


(Note: This name will be displayed publicly)