AI For The Edge: Why Open-Weight Models Matter


The rapid advancements in AI have brought powerful large language models (LLMs) to the forefront. However, most high-performing models are massive, compute-heavy, and require cloud-based inference, making them impractical for edge computing. The recent release of DeepSeek-R1 is an early, but unlikely to be the only, example of how open-weight AI models, combined with efficient distillation t... » read more

The Rise Of Generative AI On The Edge


Artificial intelligence (AI) and machine learning (ML) have undergone significant transformations over the past decade. The revolution of convolutional neural networks (CNNs) and recurrent neural networks (RNNs) is evolving toward the adoption of transformers and generative AI (GenAI), marking a pivotal shift in the field. This transition is driven by the need for more accurate, efficient, and ... » read more

Power Budgets Optimized By Managing Glitch Power


“Waste not, want not,” says the old adage, and in general, that’s good advice to live by. But in the realm of chip design, wasting power is a fact of physics. Glitch power – power that gets expended due to delays in gates and/or wires – can account for up to 40% of the power budget in advanced applications like data center servers. Even in less high-powered circuits, such as those fou... » read more

Semiconductor Test Faces Technology Shifts In The AI Era


The surge in data-rich applications shows no signs of slowing down, fueling significant evolution within the global semiconductor industry. This insatiable demand for data necessitates a comprehensive ecosystem involving sensors and systems to capture data, networks to transmit it, and storage and processing power to analyze it. Successful deployment of these applications relies on the devel... » read more

Advancing AI At The IoT Edge


In a highly connected world, there is a need for more intelligent and secure computation locally and preferably on the very devices that capture data, whether it be raw or compressed video, images, or voice. End markets continue to expect compute costs to trend down, at a time when computation demands are increasing, as is evident from recently popularized AI paradigms such as large language mo... » read more

Small Language Models: A Solution To Language Model Deployment At The Edge?


While Large Language Models (LLMs) like GPT-3 and GPT-4 have quickly become synonymous with AI, LLM mass deployments in both training and inference applications have, to date, been predominately cloud-based. This is primarily due to the sheer size of the models; the resulting processing and memory requirements often overwhelm the capabilities of edge-based systems. While the efficiency of Exped... » read more

Chip Industry Week In Review


The University of Texas at Austin’s Texas Institute for Electronics (TIE) was awarded $840 million to establish a Department of Defense microelectronics manufacturing center. This center will focus on developing advanced semiconductor microsystems to enhance U.S. defense systems. The project is part of DARPA's NGMM Program. The U.S. Dept. of Commerce announced preliminary terms with Global... » read more

Vision Is Why LLMs Matter On The Edge


Large Language Models (LLMs) have taken the world by storm since the 2017 Transformers paper, but pushing them to the edge has proved problematic. Just this year, Google had to revise its plans to roll out Gemini Nano on all new Pixel models — the down-spec’d hardware options proved unable to host the model as part of a positive user experience. But the implementation of language-focused mo... » read more

How To Successfully Deploy GenAI On Edge Devices


Generative AI (GenAI) burst onto the scene and into the public’s imagination with the launch of ChatGPT in late 2022. Users were amazed at the natural language processing chatbot’s ability to turn a short text prompt into coherent humanlike text including essays, language translations, and code examples. Technology companies – impressed with ChatGPT’s abilities – have started looking ... » read more

MIPI In Next Generation Of AI IoT Devices At The Edge


The history of data processing begins in the 1960s with centralized on-site mainframes that later evolved into distributed client servers. In the beginning of this century, centralized cloud computing became attractive and began to gain momentum, becoming one of the most popular computing tools today. In recent years however, we have seen an increase in the demand for processing at the edge or ... » read more

← Older posts