ASUS Presents MGX-Powered Data-Center Solutions

Leading full stack of AI supercomputing solutions unveiled at GTC 2024

SINGAPORE – Media OutReach Newswire – 21 March 2024 – ASUS today announced its participation at the NVIDIA GTC global AI conference, where it will showcase its solutions at booth #730. On show will be the apex of ASUS GPU server innovation, ESC NM1-E1 and ESC NM2-E1, powered by the NVIDIA MGX modular reference architecture, accelerating AI supercomputing to new heights.

To help meet the increasing demands for generative AI, ASUS uses the latest technologies from NVIDIA, including the B200 Tensor Core GPU, the GB200 Grace Blackwell Superchip, and H200 NVL, to help deliver optimized AI server solutions to boost AI adoption across a wide range of industries.

To better support enterprises in establishing their own generative AI environments, ASUS offers an extensive lineup of servers, from entry-level to high-end GPU server solutions, plus a comprehensive range of liquid-cooled rack solutions, to meet diverse workloads. Additionally, by leveraging its MLPerf expertise, the ASUS team is pursuing excellence by optimizing hardware and software for large-language-model (LLM) training and inferencing and seamlessly integrating total AI solutions to meet the demanding landscape of AI supercomputing.

Tailored AI solutions with the all-new ASUS NVIDIA MGX-powered server
The latest ASUS NVIDIA MGX-powered 2U servers, ESC NM1-E1 and ESC NM2-E1, showcase the NVIDIA GH200 Grace Hopper Superchip which offer high performance and efficiency. The NVIDIA Grace CPU includes Arm® Neoverse V9 CPU cores with Scalable Vector Extensions (SVE2) and is powered by NVIDIA NVLink-C2C technology. Integrating with NVIDIA BlueField-3 DPUs and ConnectX-7 network adapters, ASUS MGX-powered servers deliver a blazing data throughput of 400Gb/s, ideal for enterprise AI development and deployment. Coupled with NVIDIA AI Enterprise, an end-to-end, cloud-native software platform for building and deploying enterprise-grade AI applications, the MGX-powered ESC NM1-E1 provides unparalleled flexibility and scalability for AI-driven data centers, HPC, data analytics and NVIDIA Omniverse applications.

Advanced liquid-cooling technology
The surge in AI applications has heightened the demand for advanced server-cooling technology. ASUS direct-to-chip (D2C) cooling offers a quick, simple option that distinguishes itself from the competition. D2C can be rapidly deployed, lowering data center power-usage effectiveness (PUE) ratios. ASUS servers, ESC N8-E11 and RS720QN-E11-RS24U, support manifolds and cool plates, enabling diverse cooling solutions. Additionally, ASUS servers accommodate a rear-door heat exchanger compliant with standard rack-server designs, eliminating the need to replace all racks — only the rear door is required to enable liquid cooling in the rack. By closely collaborating with industry-leading cooling solution providers, ASUS provides enterprise-grade comprehensive cooling solutions and is committed to minimizing data center PUE, carbon emissions and energy consumption to assist in the design and construction of greener data centers.

Confident AI software solutions
With its world-leading expertise in AI supercomputing, ASUS provides optimized server design and rack integration for data-intensive workloads. At GTC, ASUS will showcase ESC4000A-E12 to demonstrate a no-code AI platform with an integrated software stack, enabling businesses to accelerate AI development on LLM pre-training, fine-tuning and inference — reducing risks and time-to-market without starting from scratch. Additionally, ASUS provides a comprehensive solution to support different LLM tokens from 7B, 33B and even over 180B with customized software, facilitating seamless server data dispatching. By optimizing the allocation of GPU resources for fine-tune training, the software stack ensures that AI applications and workloads can run without wasting resources, which helps to maximize efficiency and return on investment (ROI). Furthermore, the software-hardware synergy delivered by ASUS provides businesses with the flexibility to choose the AI capabilities that best fit their needs, allowing them to push ROI still further.

This innovative software approach optimizes the allocation of dedicated GPU resources for AI training and inferencing, boosting system performance. The integrated software-hardware synergy caters to diverse AI training needs, empowering businesses of all sizes, including SMBs, to leverage advanced AI capabilities with ease and efficiency.

To address the evolving requirements of enterprise IoT applications, ASUS, renowned for its robust computing capabilities, is collaborating with industrial partners, software experts and domain-focused integrators. These collaborations aim to offer turnkey server support for complete solutions, including full installation and testing for modern data-center, AI and HPC applications.

ASUS servers are available worldwide. Please visit for more ASUS data-center solutions or please contact your local ASUS representative for further information.
Hashtag: #ASUS #AI #servers

The issuer is solely responsible for the content of this announcement.

About ASUS

ASUS is a global technology leader that provides the world’s most innovative and intuitive devices, components and solutions to deliver incredible experiences that enhance the lives of people everywhere. With its team of 5,000 in-house R&D experts, ASUS is world-renowned for continuously reimagining today’s technologies for tomorrow, garners more than 11 awards every day for quality, innovation and design, and is ranked among Fortune’s World’s Most Admired Companies.