A new technical paper titled "AI and Memory Wall" was published by researchers at UC Berkeley, ICSI, and LBNL.
Abstract
"The availability of unprecedented unsupervised training data, along with neural scaling laws, has resulted in an unprecedented surge in model size and compute requirements for serving/training LLMs. However, the main performance bottleneck is increasingly shifting to memo...
» read more