AI COMPUTE

Purpose built AI compute for AI builders

An ideal supplement to your existing hyperscalers, Zai Node provides HPC GPUs, turnkey orchestration and advanced models in an integrated platform for developing and running cutting-edge AI.

OUR COMPUTE OFFERINGS

ON PREM

Buy or rent our HPC servers, at your premises for complete control and data protection. With NVIDIA RTX Pro 6000 Blackwell GPUs or custom.

Find out more >

CLOUD

Hosted in your cloud or Zai Node’s data centres in Australia and Germany. Fully managed stack that delivers access to cutting edge NVIDIA GPUs.

Find out more >

EDGE DEVICES

Proprietary, purpose-built hardware and sensors with DGX Spark, enabling real-time data processing, ML, analysis with no cloud, offline.

Find out more >

DESKTOP WORKSTATION

HPC workstation With NVIDIA RTX 6000 Blackwell graphics card for local dev to train, fine-tune and run complex models and large datasets.

Find out more >

Why have your own AI compute?

Why Zai Node AI compute?

  • Local control

    Keep data in your Australian or European regions.

  • Flexible design, easy upgrades

    Choose bare metal, managed clusters or APIs and flex anytime in 24 hours, not months.

  • Scalable results, better ROI

    Zai Node helps you scale AI while reducing cost. Connect models, agents, data and work under one lineage system.

  • Portability options

    Deploy seamlessly across cloud and/or on-prem; dedicated environments for security, compliance, or performance; or self-hosted for full control.

  • Fast access, zero config

    Achieve peace of mind when standing up AI infra with NVIDIA, app frameworks, dev tools and microservices optimised to run on Zai Node systems.

  • Expert support & warranty

    Hands-on support from dedicated engineers for any infra, OS or ML issues.

ON PREM

No compromise data protection, with our on-prem fully air-gapped HPC servers, you get the performance of the cloud with the security of your own premises. Your data, your servers, your control. Run serious workloads, train models, process data, build inference endpoints.

Buy or rent

Flexible ownership to suit your CapEx or OpEx needs.

Latest tech

Equipped with NVIDIA RTX Pro 6000 Blackwell GPUs.

Custom build options

Tailored to your workloads and budgets. Choose CPU, GPU, RAM and storage.

SERVER SPECIFICATIONS

GPU — 8x NVIDIA RTX 6000 PRO Blackwell Server Edition 96GB Graphics Card (768 VRAM), TFLOPS 1008

CPU — 2x Intel XEON 6767P 64C/128T 2.40/3.90GHz CPU; or 2x AMD EPYC 9254 (24-core, 128MB cache)

RAM — 1024GB 6400Mhz DDR5 ECC RAM; or 384GB DDR5 ECC RDIMM

OS — Ubuntu Server 22.04 LTS

Storage — 1.92TB Micron 7450 PCIe 4.0 NVMe SSD; or 2x 1TB PCIe 4.0 NVMe M.2 SSD Boot Drive

Network — Dual 10GbE ports; or 2x 25G Network Adapter

16TB PCIe 5.0 NVMe SSD Data Drive

Chassis — ASUS ESC8000A-E12P; or NVIDIA MGX 4U

Dimensions — Width: 9.5" (240mm), Height: 22.8" (580mm), Depth: 20" (560mm), Weight: 43-55 lbs

Other — N + N Redundancy; Static IP

CLOUD / DATA CENTRES

Zai Node’s servers hosted in your data centre or ours in Australia and Germany. Fully managed stack that delivers access to cutting edge NVIDIA GPUs.

Buy or rent

Flexible ownership to suit your CapEx or OpEx needs.

Tier 1 data centres

Our IRAP certified DC, dedicated private environment to maintain strict governance and compliance, with VPN, fast internet, and direct connects.

Managed infrastructure

Access to fully orchestrated GPU clusters with native integrations into existing DevOps workflows.

OUR DC AT EQUINIX ME2 IN MELBOURNE
    OUR HPC SERVERS, ON PREM
OUR SECURE AI COMPUTE RACKSPACE 

EDGE DEVICES

Real-time data processing at the rugged edge

Our proprietary edge hardware and sensors are purpose built, enabling real-time data processing, machine learning, and analysis without the cloud or internet.

CHALLENGES ZAI NODE SOLVES

Modern IOT and smart devices generate enormous volumes of sensor, video, CAN and telemetry data, but the data is:

Passive data

Collected passively and analysed later.

Needs high bandwith

Dependent on high‑bandwidth connectivity, online only.

Poor hardware

Processed on general‑purpose hardware not designed for harsh insitu conditions.

ZAI NODE’S KEY CAPABILITIES

Our solutions combine ruggedised AI hardware, industry‑grade interfaces, and embedded intelligence.

Real time

Inference at the edge immediately detects anomalies, failures and performance deviations.

On device

Reducing and prioritising data with only the most valuable data logged or transmitted.

Integrated

Native support for related networks, sensors and workflows.

Private

Processing sensitive data directly on the device, complete privacy.

Offline

Offline and low‑latency operation with data available and processed without connectivity.

IDEAL FOR MANUFACTURERS AND OPERATORS

Zai Node supports asset owners and operators, to modernise existing sensor and device fleets. We also work with sensor manufacturers to train, manage, deploy and embed AI/ML and software alongside your sensors.

Zai Node’s edge AI in production

  • Vehicle monitoring and testing

    We design and manufacture edge AI–based testing devices purpose‑built for the automotive industry. Our hardware is embedded directly into automotive prototypes, test vehicles and pre‑production platforms to deliver real‑time, AI‑driven test intel to global automotive manufacturers and Tier 1 suppliers.

    Our solution processes data at the edge, inside the vehicle or test environment. This enables immediate insight, fast iteration, and reliable operation in conditions where connectivity, latency and robustness are critical.

  • Safety, security cameras

    Zai Node’s small, rugged AI computer is installed near or inside a camera. It ingests live video streams and runs a local VLM to detect predefined conditions, e.g. near misses with vehicles or equipment, unsafe or unauthorised acts.

    Decisions happen in milliseconds (low latency), works with no connectivity, and data is processed locally.

    Our edge device includes a GPU, suitable ports, rugged design, wide temp tolerance, CAN and Ethernet interfaces.

  • Asset & machine monitoring

    We embed edge AI devices into assets to continuously assess their condition, detect early signs of failure and alert operators in real time.

    This is where our rugged edge hardware excels with always-on operation/low power; wide temperature range; dust, moisture and vibration resistance; data privacy; and multi-sensor (analog, digital, CAN, Modbus).

    Our devices can ingest data from many sensor types at once, e.g. load, vibration, temperature, water, humidity, gas, pressure, current etc.

  • Mining

    Our team has deployed edge-AI devices at the blast bench to continuously monitor drill hole condition, verify charging integrity, and detect safety risks in real time, without connectivity. This allows miners to move from manual inspection to intelligent, data-driven drill and blast operations.

    Our in-hole sensors collect and real time process distance/depth, incline, water content, temp, sound, gas/air mix, pressure and vibration.

DESKTOP WORKSTATIONS

No compromise data protection, with Zai Node’s desktop HPCs. Your data, your personal server, your control. Run serious workloads, train models, process data, build inference endpoints.

Buy or rent

Flexible ownership to suit your CapEx or OpEx needs.

Latest tech

Equipped with NVIDIA RTX Pro 6000 Blackwell GPUs.

Custom options

Tailored to your work and budget. Choose CPU, GPU, RAM and storage.

DESKTOP SPECIFICATIONS

GPU — 4x NVIDIA RTX 6000 PRO Blackwell Max-Q 96GB Graphics Card (384 VRAM) 300W per GPU

CPU — AMD Ryzen Threadripper PRO 7975WX (liquid cooled with Silverstone XE360-TR5); 32 cores / 64 threads; Base clock: 4.0 GHz, Boost up to 5.3 GHz; 8-channel DDR5 memory controller

RAM — 256GB DDR5 RAM; 8 channels

OS — Ubuntu Server

Storage — 8TB total: 4x 2TB PCIe 5.0 NVMe SSDs x4 lanes each (up to 14,900 MB/s theoretical for each NVMe module)

Network — Dual 10GbE ports; or 2x 25G Network Adapter

16TB PCIe 5.0 NVMe SSD Data Drive

Chassis — ATX

Dimensions — Width: 9.5" (240mm), Height: 12" (304mm), Depth: 20" (560mm), Weight: 43-55 lbs

Other — N + N Redundancy; Static IP

Key attributes of Zai Node’s compute

  • Power use

    Our AI servers require extreme rack densities (30–60 kW per rack). Scaling this demands advanced electrical design, intelligent distribution and redundant power systems. Our infra, chips, models and software are carefully designed to minimise energy use and maximise value per watt utilised.

  • Cooling systems

    Our closed cooling loops eliminate evaporation, drastically reducing municipal water demand. Our system design prioritises drought resilience and minimal draw from community resources. We have built a novel prototype immersion cooling GPU cluster, we believe an Australian first with significant energy efficiency benefits.

  • Networking

    Our segregated high-performance network includes hardware management, hypervisor and cluster management, inference serving networks, and training networks with extreme east–west bandwidth (400–800 Gbps per node), enabling frontier LLM training. Redundancy includes multi-path fibre and automatic failover provide resilience and continuous availability.

  • Compute & storage

    Dense GPU servers with NVLink and PCIe Gen5 architectures, optimised for AI training and inference. CPU compute includes high-core-count systems for orchestration, preprocessing and hybrid applications. Storage with hot NVMe-based distributed storage for active training datasets; and cool, erasure-coded object storage for archival and compliance requirements.

  • Simple & fast

    We can spin up and spin down GPU compute with 24 hours notice. We also engineer controls to prevent unexpected run times, when users forget to pause compute.

  • Quick cross connects

    As we’re colocated in major DCs, cross connects are fast and secure. Ideal for connecting your AI workloads to other devices, networks or DCs, all while avoiding the public internet. Connect to e.g. Anthropic, AWS Bedrock, AWS, Bard AI, OpenAI, Azure, Kubernetes Engine … all within minutes.

  • Hardware Manifest & Attestation

    For each server or workstation, Zai Node provides fully certified Hardware Manifest and Attestation documentation, ensuring security protocols are met.

  • Third party inspections

    Unlike hyperscalers, Zai Node makes our infra open for physical and digital third party inspections and verifications.

  • Certificate of Origin

    Our servers and workstations are accompanied by a Certificate of Australian Origin.

A genuine complement to hyperscale AI compute. Sovereign, private, local, low-latency, cost-effective.