From data to tokens
ASUS AI Factory delivers intelligence at scale

From data to tokens
The ASUS AI Factory brings innovation to life by transforming data into actionable intelligence and scaling ideas into real-world AI solutions. ASUS delivers comprehensive AI solutions, driving continuous innovation and maintaining a leading position. As an AI company, ASUS is committed to building an intelligent and sustainable future through design-thinking to weave AI into every aspect of our business. Our cross-product portfolio is helping lead the way to make AI accessible everywhere for everyone and make sure that everybody can benefit from the possibilities guaranteed by AI.
ASUS showcased a holistic AI ecosystem that extends beyond the PC, empowering users across diverse segments. This ecosystem is supported by robust AI infrastructure, encompassing scalable cloud and powerful on-premise solutions that seamlessly integrate hardware, software, and services, supporting the development and deployment of AI for various enterprise AI workloads. Our diverse portfolio includes intelligent solutions, strategically tailored for key segments such as developers, AI Researchers, data scientists, biotech, mission-critical applications, manufacturing, 5G and networking, and could-service providers.
The comprehensive ASUS AI Factory approach seamlessly integrates hardware, software platforms, and services to support our clients’ success in AI deployment. From edge AI to AI supercomputing, our scalable infrastructure is designed to meet the demands of diverse workloads. Critically, we help organizations accelerate AI adoption while optimizing total cost of ownership (TCO).
With a global network of over 2,200 service centers, ASUS delivers SLA-driven, tailored solutions that span the entire product lifecycle — from design and manufacturing to ongoing support. Our robust AI and data governance framework ensures data quality, regulatory compliance, and trust, safeguarding the integrity and accuracy of AI outcomes.
Discover how the ASUS AI Factory is revolutionizing applications across industries.
Driven by our All-in-AI server strategy, we work closely with strategic partners to advance the AI ecosystem. Inspired by industrial manufacturing principles, the ASUS AI Factory approach treats AI development as a streamlined, repeatable process — turning data into valuable AI-driven insights and actions efficiently.
The AI Factory with ASUS AI POD, validated by NVIDIA, provides guidance for developing, deploying, and managing agentic AI, physical AI, and HPC workloads on the NVIDIA Blackwell platform on-premises. Designed for enterprise IT, it provides accelerated computing, networking, storage, and software to help deliver faster time-to-value AI factory deployments while mitigating deployment risks.
Below are the reference architecture designs that help clients use approved practices, acting as a knowledge repository and a standardized framework for diverse applications.
Grace
Blackwell
MGX
Platform
The advanced ASUS AI POD, accelerated by NVIDIA GB200/GB300 NVL72 racks and incorporating NVIDIA Quantum InfiniBand or NVIDIA Spectrum-X Ethernet networking platforms, features liquid cooling to enable a non-blocking 576-GPU cluster across eight racks, or an air-cooled solution to support one rack with 72 GPUs. This ultra-dense, ultra-efficient architecture redefines AI reasoning computing performance and efficiency.
NVIDIA HGX™ reference architecture optimized by ASUS delivers unmatched efficiency, thermal management, and GPU density for accelerated AI fine-tuning, LLM inference, and training. Built on the ASUS XA NB3I-E12 with NVIDIA HGX™ B300 or ESC NB8-E11 embedded with NVIDIA HGX™ B200, this centralized rack solution offers unmatched manufacturing capacity for liquid-cooled or air-cooled rack systems, ensuring timely delivery, reduced total cost of ownership (TCO), and consistent performance.
B300LC | Quantum | Spectrum |
---|---|---|
GPU Support | B300 NVL8 | B300 NVL8 |
Form Factor | 5RU | 5RU |
Rack/Node | 8x Node per Rack / 9x rack | 8x Node per Rack / 4x rack |
1 SU | 9x rack / 72x node | 4x rack / 32x Node |
Switch (N/S) | 8x SN5610 (Leaf) / 4x SN5610 (Spine) | 2x SN5610 (Leaf) |
Switch (E/W) | 8x Q3400-RA (Leaf) / 4x Q3400-RA (Spine) | 8x SN5610 (Leaf) / 4x SN5610 (Spine) |
Server | XA NB3I-E12L | XA NB3I-E12L |
Storage | RS501A-E12-RS12U | RS501A-E12-RS12U |
MGMT | RS700-E12-RS4U | RS700-E12-RS4U |
B300AC | Quantum | Spectrum |
---|---|---|
GPU Support | B300 NVL8 | B300 NVL8 |
Form Factor | 9RU | 9RU |
Rack/Node | 4x Node per Rack / 18x rack | 4x Node per Rack / 8x rack |
1 SU | 18x rack / 72x node | 8x rack / 32x Node |
Switch (N/S) | 8x SN5610 (Leaf) / 4x SN5610 (Spine) | 2x SN5610 (Leaf) |
Switch (E/W) | 8x Q3400-RA (Leaf) / 4x Q3400-RA (Spine) | 8x SN5610 (Leaf) / 4x SN5610 (Spine) |
Server | XA NB3I-E12 | XA NB3I-E12 |
Storage | RS501A-E12-RS12U | RS501A-E12-RS12U |
MGMT | RS700-E12-RS4U | RS700-E12-RS4U |
B200AC | Quantum | Spectrum |
---|---|---|
GPU Support | B200 NVL8 | B200 NVL8 |
Form Factor | 10RU | 10RU |
Rack/Node | 4x Node per Rack / 8x rack | 4x Node per Rack / 8x rack |
1 SU | 8x rack / 32x node | 8x rack / 32x Node |
Switch (N/S) | 2x SN5610 (Leaf) | 2x SN5610 (Leaf) |
Switch (E/W) | 8x QM9700 (Leaf) / 4x QM9700 (Spine) | 4x SN5610 (Leaf) / 2x SN5610 (Spine) |
Server | ESC N8B-E11 | ESC N8B-E11 |
Storage | RS501A-E12-RS12U | RS501A-E12-RS12U |
MGMT | RS700-E12-RS4U | RS700-E12-RS4U |
ASUS presents NVIDIA MGX-compliant rack designs with the ASUS ESC8000 series, featuring dual Intel® Xeon® 6 processors and RTX PRO™ 6000 Blackwell Server Edition with the latest NVIDIA ConnectX-8 SuperNIC – supporting speeds of up to 800Gb/s or other scalable configurations – delivering exceptional expandability and performance for state-of-the-art AI workloads. Integration with the NVIDIA AI Enterprise software platform provides highly-scalable, full-stack server solutions that meet the demanding requirements of modern computing.
MGX CX8 | Quantum | Spectrum |
---|---|---|
GPU Support | H200 NVL, L40S, RTX Pro 6000 BSE | H200 NVL, L40S, RTX Pro 6000 BSE |
Form Factor | 4RU | 4RU |
Rack/Node | 8x Node per Rack / 4x rack in total or 4x Node per Rack / 8x rack in total |
8x Node per Rack / 4x rack in total or 4x Node per Rack / 8x rack in total |
1 SU | 8x rack / 4x or 8x rack / 32x Node | 8x rack / 4x or 8x rack / 32x Node |
Switch (N/S) | 2x SN5610 (Leaf) | 2x SN5610 (Leaf) |
Switch (E/W) | 4x QM9700 (Leaf) / 2x QM9700 (Spine) | 2x SN5610 (Leaf) |
Server | ESC8000-E12P ESC8000A-E13P |
ESC8000-E12P ESC8000A-E13P |
Storage | RS501A-E12-RS12U | RS501A-E12-RS12U |
MGMT | RS700-E12-RS4U | RS700-E12-RS4U |
ASUS is committed to not only enabling rapid deployment of your data center but also ensuring its smooth and reliable operation. That’s why we offer a comprehensive suite of software tools and platforms, allowing for one-click deployment and customizable interfaces and services—such as billing—tailored to your specific business needs.
Automated large-scale deployment and provisioning
Full-stack visibility and proactive alerts
Centralize server management
Smart fabric monitoring and network optimization
Scalable AI training and inference platform
Efficient compute resource and job orchestration
As your trusted NVIDIA® Omniverse integration provider, ASUS offer a complete solution from strategic consulting to deployment support to help you unlock the full potential of industrial digitalization.
GB200 NVL72 | Quantum | ||
---|---|---|---|
Quantum | Quantum | Spectrum | |
GPU Support | GB200 | GB300 | GB300 |
Form Factor | 48RU | 48RU | 48RU |
Rack/Node | 18x Compute Trays / 9x Switch Trays | 18x Compute Trays / 9x Switch Trays | 18x Compute Trays / 9x Switch Trays |
1 SU | 16x rack / 288x node | 16x rack / 288x node | 16x rack / 288x node |
Switch (N/S) | 8x SN5610 (Leaf) / 4x SN5610 (Spine) / 4x SN5610 (Core) | 8x SN5610 (Leaf) / 4x SN5610 (Spine) | 8x SN5610 (Leaf) / 4x SN5610 (Spine) |
Switch (E/W) | 16x Q9700 (Leaf) / 18x Q9700 (Spine) | 16x Q3400-RA (Leaf) / 9x Q3400-RA (Spine) | 16x Q3400-RA (Leaf) / 9x Q3400-RA (Spine) |
Server | ESC NM2N721-E1 | XA GB721-E2 | XA GB721-E2 |
Storage | RS501A-E12-RS12U | RS501A-E12-RS12U | RS501A-E12-RS12U |
MGMT | RS700-E12-RS4U | RS700-E12-RS4U | RS700-E12-RS4U |
GB Series | B Series | MGX/HGX Series | |
---|---|---|---|
Server Model | ASUS RS501A-E12-RS12U | ||
CPU & Memory | AMD 64C ( 9005 series ) & 768G RAM | ||
Network | 400G(200x2) IB NIC x 2 | ||
Storage | Boot disk : 2 x 960G NVMe M.2 Storage disk : 16 x 30.72 TLC NVMe U.2 |
||
Server nodes | 27 | 19 | 8 |
Usable Capacity | 10PB | 7PB | 2.2PB |
Read Speed | 2284GB | 1607GB | 676GB |
Write Speed | 691GB | 486GB | 164GB |
Read IOPS | 49.5 Million | 34.8 Million | 14.6 Million |
Write IOPS | 12.4 Million | 8.7 Million | 3.6 Million |
Stripe Width | 16+2+1 | 16+2+1 | 5+2+1 |
Certified for WEKA Data Platform
Efficient and high-bandwidth network connectivity is crucial for NVIDIA GB200 NVL72 performance. Below is an outline of the typical external network architecture and technologies used to connect multiple NVL72 systems and integrate them into a larger network infrastructure.
ODs | 1 | 1 | 2 | 4 | |||
Racks | 1 | 2 | 4 | 8 | 6 | 32 | 64 |
Computing Tray | 18 | 36 | 72 | 144 | 288 | 576 | 1152 |
Blackwell GPU | 72 | 144 | 288 | 576 | 1152 | 2304 | 4608 |
Bluefield-3 (DPU) | 32 | 72 | 144 | 288 | 576 | 1152 | 2304 |
Switch | |||||||
# of Spine Switches | 0 | 2 | 4 | 8 | 16 | 32 | |
# of Leaf Switches | 2 | 4 | 8 | 16 | 32 | 64 |
ASUS, in collaboration with partners, offers comprehensive cabinet-level liquid-cooling solutions. These include CPU/GPU cold plates, cooling distribution units and cooling towers, all designed to minimize power consumption and optimize power-usage effectiveness (PUE) in data centers.
Ideal for small-scale data centers with compact facilities.
Designed to meet the needs of existing air-cooled data centers and easily integrate with current
infrastructure.
Perfect for enterprises seeking immediate implementation and deployment
Ideal for large-scale, extensive infrastructure with high
workloads.
Provides long-term, low PUE with sustained energy
efficiency over time.
Reduces TCO for maximum value and cost-effective operations.