ATOM™-Max

Boosted Performance for Large-Scale Inference

Description

Boosted Performance for Large-Scale Inference

ATOM™-Max is the best in-class AI accelerator engineered for data centers and enterprises, delivering an impressive performance boost for large-scale AI inference workloads. ATOM™-Max achieves 128 TFLOPS (FP16) and up to 512 TOPS (INT8) / 1024 TOPS (INT4), with a bandwidth of 1024 GB/s, ensuring maximum throughput for demanding applications.

Key features

Massive Performance Boost

Exceptional AI compute with a substantial bandwidth boost

Direct Card-to-Card Communication

Seamless and fast data exchange using PCIe Gen5 x16, reducing latency and enhancing scalability

Optimized for Large-Scale Inference

Specifically designed to meet the needs of large-scale enterprises and AI-driven infrastructures

Product specs

FP16
128 TFLOPS
INT8 / INT4
512 TOPS / 1024 TOPS
Input Power
DC 12V (CPU 8-pin power connector)
Max Power Consumption
350 W
Thermal
Air Cooling (passive)
Memory
GDDR6 64 GB, 1024 GB/s
Host Interface
PCIe Gen5 x16, 64 GB/s
Intercard Support
Yes
Form Factor
FHFL Dual Slot

ATOM™-Max Now in Mass Production

ATOM™ Series

ATOM™

Cost-efficient, Powerful AI Acceleration for Small-sized Data centers

Explore ATOM™

ATOM™-Lite

Low-power, Yet Highly Powerful AI Inference at the Edge

Explore ATOM™-Lite

ATOM™ SoC

Inference AI Accelerator for Data centers

ATOM™ is a fast and power-efficient System-on-Chip for AI inference with remarkably low latency, conceived for deployment in data centers or cloud service providers.
With international recognition from the industry-standard benchmark MLperf™ v3.0, ATOM™ can be scaled to accelerate state-of-the-art AI models of various sizes.
Explore ATOM™ SoC