Qualcomm has announced some details of its tilt at the AI datacenter market by revealing a pair of accelerators and rack scale systems to house them, all focused on inferencing workloads.
The company offered scant technical details about its new AI200 and AI250 “chip-based accelerator cards”, saying only that the AI 200 supports 768 GB of LPDDR memory per card, and the AI250 will offer “innovative memory architecture based on near-memory computing” and represent “a generational leap in efficiency and performance for AI inference workloads by delivering greater than 10x higher effective memory bandwidth and much lower power consumption.”
Qualcomm will ship the cards in pre-configured racks that will use “direct liquid cooling for thermal efficiency, PCIe for scale up, Ethernet for scale o

The Register

The radio station 99.5 The Apple
WRCB-TV
NBC News
Raw Story
AlterNet
Reuters US Domestic
The Daily Beast
CNA Entertainment