Qualcomm's AI200 and AI250: Revolutionizing the data center AI landscape

Qualcomm unveils its cutting-edge AI200 and AI250 solutions, enhancing generative AI capabilities in data centres with remarkable efficiency and scalability.

  • Monday, 27th October 2025 Posted 5 months ago in by Aaron Sandhu

Qualcomm Technologies, Inc. has announced the launch of its latest AI-optimised solutions for data centres: the AI200 and AI250 accelerator cards and racks. Building on its NPU technology leadership, these offerings bring unrivalled rack-scale performance and memory capacity, setting a new standard for generative AI inference.

The Qualcomm AI200 is tailored to deliver low total cost of ownership and optimised performance for large language models and other AI workloads. With support for 768 GB of LPDDR per card, it addresses high memory needs with an emphasis on scalability and flexibility for AI tasks.

Meanwhile, the Qualcomm AI250 introduces a pioneering memory architecture based on near-memory computing. This delivers a more than 10x improvement in memory bandwidth efficiency and significantly reduces power consumption, facilitating the robust application of AI in hardware tailored to meet diverse customer needs.

Both solutions are equipped with features like direct liquid cooling, PCIe for scalability, and Ethernet for expansion. They also incorporate confidential computing to safeguard AI processes, with a power draw of 160 kW per rack. This allows Qualcomm's offerings to meet the demands of modern data centre environments.

According to Durga Malladi, SVP & GM at Qualcomm, "Our rich software stack and open ecosystem support  make it easier than ever for developers and enterprises to integrate, manage, and scale already trained AI models on our optimised AI inference solutions. With seamless compatibility for leading AI frameworks and one-click model deployment, Qualcomm AI200 and AI250 are designed for frictionless adoption and rapid innovation."

The combined software stack supports a wide array of machine learning frameworks and inference engines, optimising for generative AI tasks with techniques like disaggregated serving. Developers benefit from streamlined model integration and access to AI applications, libraries, and tools, ensuring operational efficiency.

Anticipated to be commercially available from 2026 and 2027, the AI200 and AI250 embody Qualcomm's commitment to annual progress in data centre AI, focusing on performance, energy efficiency, and cost-effectiveness.

Jacobs introduces a digital twin solution for AI data centres, aiming to enhance planning,...
Corning partners with US Conec to enhance AI network connectivity through PRIZM TMT technology,...
TE Connectivity introduces its 56G MezzaWave connectors, offering enhanced signal integrity for...
Dell expands its security measures to help protect organisations against quantum computing and AI...
Exploring Keysight's new solution for error performance validation in AI-focused data centres,...
Stellanor appoints Howard Pheby as Chief Commercial Officer to lead commercial operations during...
Antin Infrastructure Partners has acquired NorthC Datacenters, adding to its portfolio of European...
Phononic has expanded its solid-state cooling solutions for AI and high-performance data centres,...