Weighing efficiency gains vs. the scale of personalization.
By Ramin Farjadrad and Syrus Ziai
There is a big push to achieve greater scale, performance and sustainability to fuel the AI revolution. More speed, more memory bandwidth, less power — these are the holy grails.
Naturally, the one-two punch of StarGate and DeepSeek last week has raised many questions in our ecosystem and with our various stakeholders. Can DeepSeek be real? And if so, what does this mean for investments in AI infrastructure like StarGate?
DeepSeek’s apparent breakthrough in optimizing AI training highlights an undoubtedly major shift. If the claims are true, we can now achieve powerful AI capabilities with significantly reduced computational costs. If model sizes could indeed be shrunk to cut training complexity by 10X or more, this would make AI more accessible beyond traditional data centers—fueling AI’s expansion into Edge devices and everyday applications.
DeepSeek leverages algorithms such as Mixture of Experts (MoE), which demand a lot of memory bandwidth and produce large amounts of temporary output token, which need to be stored in memory and read back. Eliyan’s die-to-memory technology with significantly higher bandwidth/power can directly address this market expansion.
As history has shown, once new technology becomes more accessible to the masses, applications and use cases explode. We have seen this story before with the birth and rise of the Internet. The technologies that were created over the past two decades helped provide higher bandwidth Internet at lower costs, which helped dramatically expand the Internet market and the market cap of the companies delivering valuable products and services in that space, and to the overall economy. This expansion created many markets that did not exist before, and with that the demand for enabling hardware and software technologies exploded.
The next question is if we completely discount the impact of this dynamic in expanding the market, does this mean hyperscaler AI infrastructure spending will slow? Far from it.
The key distinction lies in scale—not just in model size, but in the sheer volume of data that AI must process for hyper-personalization.
Let’s compare:
The difference in scale is staggering:
This shift from training AI on static, finite knowledge to dynamic, ever-expanding personal data is why hyperscalers will continue to invest heavily in AI infrastructure. While DeepSeek’s reported optimizations would make AI more efficient, the push for ultra-personalized AI requires orders of magnitude more compute power, storage, and high-performance AI hardware.
That’s why Hyperscalers, Open AI, StarGate and various other initiatives and entities are spending hundreds of billions of dollars on AI infrastructure, knowing well that they may not achieve positive ROI just by offering customers AI assistant apps or tools (this is especially true as the price of such AI assistant tools will surely drop significantly once the models that offer such capability become open source and more than two or three companies offer the same capabilities).
So while DeepSeek may reduce the compute hardware for the training by an order of magnitude, and allow training based on the existing knowledge base to be much faster (reducing from trillions of parameters to <100 billion), the characteristic dataset of all billions of humans online can be in hundreds to several thousands of trillions of parameters, and AI hardware is still far from being capable of training at that scale.
The bottom line? Efficiency gains won’t slow AI hardware demand—they will accelerate AI’s ubiquity while fueling an even larger demand for computational power. We are just at the beginning of an era where AI models aren’t just trained on what humanity has already discovered — but on who we are, individually, in real time.
Syrus Ziai is a co-founder of Eliyan.
Leave a Reply