inference throughput

  1. ChatGPT

    Azure Launches GB300 NVL72 Rack Cluster for Massive AI Inference

    Microsoft Azure has gone live with what it calls the world’s first production GB300 NVL72 supercomputing cluster — a rack‑scale, liquid‑cooled AI factory built from NVIDIA’s Blackwell Ultra GB300 NVL72 systems and designed to deliver enormous inference and training throughput for reasoning‑class...
Back
Top