SuperMicro Unveils Super Clusters Powered by NVIDIA Chips

<p><strong>San Jose and Taipei&comma; Taiwan – <&sol;strong>Supermicro is introducing at COMPUTEX in Taiwan a ready-to-deploy liquid-cooled AI data center&comma; designed for cloud-native solutions that accelerate generative AI adoption for enterprises across industries with its SuperClusters&period; The servers are optimized for the <a href&equals;"https&colon;&sol;&sol;www&period;nvidia&period;com&sol;en-us&sol;data-center&sol;products&sol;ai-enterprise&sol;">NVIDIA AI Enterprise<&sol;a> software platform for the development and deployment of generative AI&period; With Supermicro&&num;8217&semi;s 4U liquid-cooled&comma; NVIDIA recently introduced <a href&equals;"https&colon;&sol;&sol;www&period;nvidia&period;com&sol;en-us&sol;data-center&sol;technologies&sol;blackwell-architecture&sol;">Blackwell GPU<&sol;a>s can fully unleash 20 PetaFLOPS on a single GPU of AI performance and demonstrate 4X better AI training and 30X better inference performance than the previous GPUs with additional cost savings&period;<&sol;p>&NewLine;<p>Supermicro recently announced a complete line of NVIDIA Blackwell architecture-based products for the new <a href&equals;"https&colon;&sol;&sol;nvidianews&period;nvidia&period;com&sol;news&sol;nvidia-blackwell-platform-arrives-to-power-a-new-era-of-computing">NVIDIA HGXTM B100&comma; B200&comma; and GB200 Grace Blackwell Superchip<&sol;a>&period;<&sol;p>&NewLine;<p>&&num;8220&semi;Supermicro continues to lead the industry in creating and deploying AI solutions with rack-scale liquid-cooling&comma;&&num;8221&semi; said Charles Liang&comma; president and CEO of Supermicro&period; &&num;8220&semi;Data centers with liquid-cooling can be virtually free and provide a bonus value for customers&comma; with the ongoing reduction in electricity usage&period; Our solutions are optimized with NVIDIA AI Enterprise software for customers across industries&comma; and we deliver global manufacturing capacity with world-class efficiency&period; The result is that we can reduce the time to delivery of our liquid-cooled or air-cooled turnkey clusters with NVIDIA HGX H100 and H200&comma; as well as the upcoming B100&comma; B200&comma; and GB200 solutions&period; From cold plates to CDUs to cooling towers&comma; our rack-scale total liquid cooling solutions can reduce ongoing data center power usage by up to 40&percnt;&period;&&num;8221&semi;<&sol;p>&NewLine;<p>CEO Liang is a Taiwan native&period;<&sol;p>&NewLine;<p>At COMPUTEX 2024&comma; Supermicro is revealing its upcoming systems optimized for the NVIDIA Blackwell GPU&comma; including a 10U air-cooled and a 4U liquid-cooled NVIDIA HGX B200-based system&period; In addition&comma; Supermicro will be offering an 8U air-cooled NVIDIA HGX B100 system and Supermicro&&num;8217&semi;s NVIDIA GB200 NVL72 rack containing 72 interconnected GPUs with <a href&equals;"https&colon;&sol;&sol;www&period;nvidia&period;com&sol;en-us&sol;data-center&sol;nvlink&sol;">NVIDIA NVLink Switches<&sol;a>&comma; as well as the new NVIDIA MGX™ systems supporting NVIDIA H200 NVL PCIe GPUs and the newly announced NVIDIA GB200 NVL2 architecture&period;<&sol;p>&NewLine;<p>&&num;8220&semi;Generative AI is driving a reset of the entire computing stack — new data centers will be GPU-accelerated and optimized for AI&comma;&&num;8221&semi; said Jensen Huang&comma; founder and CEO of NVIDIA&period; &&num;8220&semi;Supermicro has designed cutting-edge NVIDIA accelerated computing and networking solutions&comma; enabling the trillion-dollar global data centers to be optimized for the era of AI&period;&&num;8221&semi;<&sol;p>&NewLine;<p>The rapid development of large language models and the continuous new introductions of open-source models such as Meta&&num;8217&semi;s Llama-3 and Mistral&&num;8217&semi;s Mixtral 8x22B make today&&num;8217&semi;s state-of-the-art AI models more accessible for enterprises&period; The need to simplify the AI infrastructure and provide accessibility in the most cost-efficient way is paramount to supporting the current breakneck speed of the AI revolution&period; The Supermicro cloud-native AI SuperCluster bridges the gap between cloud convenience of instant access and portability&comma; leveraging the NVIDIA AI Enterprise&comma; allowing moving AI projects from pilot to production seamlessly at any scale&period; This provides the flexibility to run anywhere with securely managed data&comma; including self-hosted systems or on-premises large data centers&period;<&sol;p>&NewLine;<p>With enterprises across industries rapidly experimenting with generative AI use cases&comma; Supermicro collaborates closely with NVIDIA to ensure a seamless and flexible transition from experimentation and piloting AI applications to production deployment and large-scale data center AI&period; This result is achieved through rack and cluster-level optimization with the NVIDIA AI Enterprise software platform&comma; enabling a smooth journey from initial exploration to scalable AI implementation&period;<&sol;p>&NewLine;<p>Managed services compromise infrastructure choices&comma; data sharing&comma; and generative AI strategy control&period; NVIDIA NIM microservices&comma; part of NVIDIA AI Enterprise&comma; offer managed generative AI and open-source deployment benefits without drawbacks&period; Its versatile inference runtime with microservices accelerates generative AI deployment across a wide range of models&comma; from open-source to NVIDIA&&num;8217&semi;s foundation models&period; In addition&comma; NVIDIA NeMoTM enables custom model development with data curation&comma; advanced customization&comma; and retrieval-augmented generation &lpar;RAG&rpar; for enterprise-ready solutions&period; Combined with Supermicro&&num;8217&semi;s NVIDIA AI Enterprise ready SuperClusters&comma; NVIDIA NIM provides the fastest path to scalable&comma; accelerated Generative AI production deployments&period;<&sol;p>&NewLine;<p>Supermicro&&num;8217&semi;s current generative AI SuperCluster offerings include&colon;<&sol;p>&NewLine;<ul>&NewLine;<li><strong>Liquid-cooled Supermicro NVIDIA HGX H100&sol;H200 SuperCluster<&sol;strong> with 256 H100&sol;H200 GPUs as a scalable unit of compute in 5 racks &lpar;including 1 dedicated networking rack&rpar;<&sol;li>&NewLine;<li><strong>Air-cooled Supermicro NVIDIA HGX H100&sol;H200 SuperCluster<&sol;strong> with 256 HGX H100&sol;H200 GPUs as a scalable unit of compute in 9 racks &lpar;including 1 dedicated networking rack&rpar;<&sol;li>&NewLine;<li><strong>Supermicro NVIDIA MGX GH200 SuperCluster<&sol;strong> with 256 GH200 GraceTM Hopper Superchips as a scalable unit of compute in 9 racks &lpar;including 1 dedicated networking rack&rpar;<&sol;li>&NewLine;<&sol;ul>&NewLine;<p>Supermicro SuperClusters are NVIDIA AI Enterprise ready with NVIDIA NIM microservices and NVIDIA NeMo platform for end-to-end generative AI customization and optimized for NVIDIA Quantum-2 InfiniBand as well as the new NVIDIA Spectrum-X Ethernet platform with 400Gb&sol;s of networking speed per GPU for scaling out to a large cluster with tens of thousands of GPUs&period;<&sol;p>&NewLine;<p>Supermicro&&num;8217&semi;s upcoming SuperCluster offerings include&colon;<&sol;p>&NewLine;<ul>&NewLine;<li><strong>Supermicro NVIDIA HGX B200 SuperCluster<&sol;strong>&comma; liquid-cooled<&sol;li>&NewLine;<li><strong>Supermicro NVIDIA HGX B100&sol;B200 SuperCluster<&sol;strong>&comma; air-cooled<&sol;li>&NewLine;<li><strong>Supermicro NVIDIA GB200 NVL72 or NVL36 SuperCluster<&sol;strong>&comma; liquid-cooled<&sol;li>&NewLine;<&sol;ul>&NewLine;<p>Supermicro&&num;8217&semi;s SuperCluster solutions are optimized for LLM training&comma; deep learning&comma; and high volume and batch size inference&period; Supermicro&&num;8217&semi;s L11 and L12 validation testing and on-site deployment service provide customers with a seamless experience&period; Customers receive plug-and-play scalable units for easy deployment in a data center and faster time to results&period;<&sol;p>&NewLine;

Editor

Wispr Scores $25 Million Series A Extension

SAN FRANCISCO -- Wispr, the voice-to-text AI that turns speech into clear, polished writing in every…

1 day

Numeric Dials Up $51 Million Series B

SAN FRANCISCO -- Numeric, an AI accounting automation platform, has raised a $51 million Series…

1 day

Apple Names 45 Finalists for App Store of the Year Awards

Apple has announced 45 finalists for this year’s App Store Awards, recognizing the best apps…

2 days

UC Reaches Agreement With Nurses, Strike Canceled

The University of California (UC) and the California Nurses Association (CNA) have reached a tentative…

4 days

HouseRX Rakes In $55 Million Series B

SAN FRANCISCO -- House Rx, a health tech company focused on making specialty medications more accessible and…

5 days

King Charles Honors NVIDIA’s Jensen Huang

Britain's King has given an award to the King of NVIDIA! NVIDIA founder and CEO…

5 days