ATOM™-Max Pod

Rack-Scale AI Infrastructure with RDMA-Based High-Speed Networking

High-Performance AI Cluster with
Rack-to-Rack Scalability and RDMA Networking

Built for large-scale AI inference, the ATOM™-Max Pod is a rack-scale infrastructure designed for distributed workloads. It combines Rebellions’ AI accelerators with 400–800G RDMA-based high-speed networking and a familiar software stack—all delivered as a turnkey solution. Starting from an 8-server Mini Pod, the system scales flexibly to meet enterprise-level AI demands.

Multi-Node

System Architecture
Starting from 8 Nodes

800GB/s

RDMA Fabric
Super-Low Latency Clustering

64 NPUs

Per Rack
Linear Scaling Performance

All-in-One

Turnkey Solution
Available for Deployment

Limitless

Scale-Out Architecture

From an 8-server Mini Pod to a multi-rack cluster, ATOM™-Max Pods scale seamlessly via Rebellions Scalable Design (RSD), offering linear performance gains and flexible resource management as workloads grow.

Ultra-Low Latency

RDMA Fabric

ATOM™-Max servers within a Pod are interconnected via 400–800G RDMA networking, ensuring ultra-fast, low-latency distributed processing for resource-intensive AI models.

All-in-One

Turnkey Infrastructure

Each Pod is a fully integrated infrastructure stack—combining AI accelerators, RDMA switches, and high-speed inter-node networking—ready for immediate deployment in production environments. Say goodbye to infrastructure complexity and maximize operational efficiency.

Ready-to-Deploy

Rebellions Enterprise AI Solution

For enterprise environments, Rebellions’ Enterprise AI Solution can be integrated into the Pod, delivering full-lifecycle AI serving capabilities with cost efficiency. A turnkey solution, ready to power your AI services.

Applications

Enterprise

Streamline enterprise-wide AI adoption—from development to deployment—with scalable AI infrastructure

Construction

Enable proactive safety monitoring on construction sites with AI-powered surveillance

Healthcare

Support the AI healthcare ecosystem, from personalized wellness to precision medicine

Finance

Build next-gen financial services with secure, real-time AI processing of financial data

Manufacturing

Boost manufacturing productivity with Physical AI-powered smart factories

Telecom

Power advanced telecom services and elevate customer experience with reliable large-scale AI operations.

RBLN SDK
Deploy with Confidence from Day One.

Purpose-built for PyTorch.
Tuned for production.

High-QPS vLLM serving.
Ready out of the box.

Full Triton access with dev tools
you’ll actually use.

One-click deployment.
Zero guesswork.

Driver SDK

Core system software and essential tools for NPU excution

Firmware Kernel Driver User Model Driver System Management Tool

NPU SDK

Developer tools for model and service development

Compiler, Runtime, Profiler Hugging Face Leading Serving Frameworks (vLLM, TorchServe, Triton Inference Server, etc.)

Model Zoo

300+ ready-to-use PyTorch and TensorFlow models optimized for Rebellions NPUs

Natural Language Processing Generative AI Speech Processing Computer Vision

Cloud SDK

Cloud-based tools for managing NPU resources at scale

K8s Device Plugin Metric-Exporter Node Feature Discovery Device Installer VFIO Manager K8s Operator