Micron’s senior vice president, Jeremy Werner, told The Circuit Podcast that memory has become a strategic bottleneck for data-center inference, warning that insufficient memory can sharply cut GPU utilization while faster, larger memory can theoretically multiply the compute extracted from GPUs. The remarks underscore how storage and memory design could limit AI deployment.
Memory bottlenecks threaten data-center GPU efficiency as AI inference scales, says Micron SVP
11
May