Archives

WEKA Launches AI Storage Cluster with NVIDIA Grace CPUs

WEKA

WEKA , the AI-native data platform company, previewed the industry’s first high-performance storage solution for the  NVIDIA Grace™ CPU Superchip . The solution runs on a powerful new storage server from Supermicro powered by WEKA® Data  Platform software and Arm® Neoverse  ™ V2 cores utilizing the NVIDIA Grace CPU Superchip  and NVIDIA ConnectX-7 and NVIDIA BlueField-3 networking to accelerate enterprise AI workloads with unmatched performance density and energy efficiency.

Fueling Next-Generation AI Innovation   
Today’s AI and high-performance computing (HPC) workloads demand blazing-fast data access, but most data centers face increasing space and power constraints.

NVIDIA Grace integrates the performance of a flagship x86-64 two-socket workstation or server platform into a single module. Grace CPU Superchips are powered by 144 high-performance Arm Neoverse V2 cores, which are 2x the power efficient of traditional x86 servers. NVIDIA ConnectX-7 NICs and BlueField-3 SuperNICs feature purpose-built RDMA/RoCE acceleration and deliver high-throughput, low-latency network connectivity at speeds of up to 400Gb/s. The combination of the WEKA Data Platform’s revolutionary zero-copy software architecture on the Supermicro Petascale storage server minimizes I/O bottlenecks and reduces AI pipeline latency. This greatly optimizes GPU utilization and also accelerates AI model training and inference, dramatically improving time to first token, discovery, and insight while reducing power consumption and associated costs.

Also Read: Pure Storage Launches GenAI Pod to Boost AI Innovation 

The main benefits of the solution are:

  • Extreme Speed ​​and Scalability for Enterprise AI:  The NVIDIA Grace CPU Superchip, featuring 144 powerful Arm® Neoverse  ™ V2 cores connected by a high-performance, custom-designed NVIDIA Scalable Coherency Fabric, delivers the performance of a dual-socket x86 CPU server with half the power. NVIDIA ConnectX-7 NICs and NVIDIA BlueField-3 SuperNICs provide high-performance networking, essential for enterprise AI workloads. Combined with the WEKA Data Platform’s AI-native architecture, which accelerates time to first token by up to 10x, the solution enables optimal performance in AI data pipelines at virtually any scale.
  • Optimal Resource Utilization:  The powerful WEKA Data Platform, combined with the LPDDR5X memory architecture of the Grace CPUs, enables up to 1TB/s memory bandwidth and seamless data flow, eliminating bottlenecks. By integrating WEKA’s distributed architecture and kernel bypass technology, organizations can achieve faster AI model training, shorter epoch times, and higher inference speeds, making it the ideal solution to efficiently scale AI workloads.
  • Exceptional Power and Space Efficiency:  The WEKA Data Platform delivers 10-50x increased GPU stack efficiency to seamlessly handle large-scale AI and HPC workloads. With data copy reduction and cloud elasticity, the WEKA Platform can reduce the data infrastructure footprint by 4-7x and reduce carbon emissions, avoiding the storage of up to 260 tons of CO2e per PB per year and reducing energy costs by 10x. With the Grace CPU Superchip’s doubled energy efficiency compared to leading x86 servers, customers can do more with less, achieving sustainability goals while improving AI performance.

“AI is transforming the way businesses around the world innovate, create, and operate, but the surge in its adoption has dramatically increased data center energy consumption, which is expected to double by 2026 according to the International Atomic Energy Agency,” said Nilesh Patel , chief product officer at WEKA. “WEKA is excited to collaborate with NVIDIA, Arm, and Supermicro to develop high-performance, energy-efficient solutions for the next generation of data centers that power enterprise AI and high-performance workloads while accelerating processing of massive data and reducing time to actionable insights.”

“WEKA has teamed up with Supermicro to develop a high-performance storage solution that seamlessly integrates with the NVIDIA Grace CPU Superchip to improve the efficiency of large-scale, data-intensive AI workloads. The solution provides fast data access while reducing power consumption, enabling data-driven organizations to significantly boost their AI infrastructure,” said Ivan Goldwasser , director of datacenter CPUs at NVIDIA.

SOURCE: PRNewswire