In the deep learning inferencing game, there are plenty of chipmakers, large and small, developing custom-built ASICs aimed at this application set. But one obscure company appears to have beat them ...
Qualcomm’s AI200 and AI250 move beyond GPU-style training hardware to optimize for inference workloads, offering 10X higher memory bandwidth and reduced energy use. It’s becoming increasingly clear ...