Ways to Get Started With NVIDIA DGX Cloud

Discover the cloud-first way to get the best of NVIDIA AI.

Try NVIDIA DGX Cloud Now

Explore NVIDIA NIM™ microservices on build.nvidia.com, a free API catalog for testing, prototyping, and developing generative AI apps with fully managed, accelerated endpoints and NVIDIA Blueprints—all accelerated by NVIDIA DGX™ Cloud.

Use NVIDIA DGX Cloud Serverless Inference

Easily package and deploy inference pipelines or data preprocessing workflows in containers optimized for NVIDIA GPUs, without worrying about underlying infrastructure.

Explore NVIDIA DGX Cloud Create

Learn more about NVIDIA DGX Cloud Create, which provides optimized accelerated computing clusters on any leading cloud service provider with best-of-breed software and access to the leaders in AI innovation.

Deploy Globally with NVIDIA DGX Cloud Lepton

Get connected to the virtual global AI factory. This platform provides a unified experience to discover, procure, and utilize GPU resources for development to deployment lifecycle across multiple clouds.

Request NVIDIA DGX Cloud With NVIDIA GB200

Fuel next-gen AI breakthroughs on NVIDIA DGX Cloud with NVIDIA GB200, featuring the powerful NVIDIA Blackwell architecture and high-bandwidth NVIDIA NVLink™. NVIDIA Blackwell provides unparalleled performance, efficiency, and scale to supercharge AI model training.

FAQs

NVIDIA DGX Cloud is a unified AI platform on leading clouds to optimize performance with software, services, and AI expertise for evolving workloads. Every layer of DGX Cloud is optimized and managed by NVIDIA, ensuring highest performance of NVIDIA in the cloud. With a suite of fully managed platforms and services, DGX Cloud empowers every organization to bring AI workloads from develop to deploy, in the era of agentic and physical AI. NVIDIA DGX Cloud includes NVIDIA NeMo™ Curator on DGX Cloud, DGX Cloud Create, DGX Cloud Serverless Inference, DGX Cloud Lepton, and DGX Cloud Benchmarking.

  • Use DGX Cloud Create on leading clouds, for pretraining and fine-tuning on large-scale clusters with flexible term lengths.
  • Use DGX Cloud Serverless Inference for production-grade inference with auto-scaling, cost-efficient GPU utilization, and multi-cloud flexibility.
  • Use DGX Cloud Lepton to develop and deploy AI workloads across multiple GPU clouds, with access to a built-in marketplace for discovering and procuring compute.
  • Try DGX Cloud Benchmarking tools and recipes and follow evolving AI workload optimizations and workload-specific recipes to maximize your AI infrastructure performance.
  • Try NVIDIA NeMo Curator on DGX Cloud for large-scale video curation.

NVIDIA NIM™ provides prebuilt, optimized inference microservices that let you deploy the latest AI foundation models with security and stability on any NVIDIA-accelerated infrastructure — cloud, data center, and workstation. You can experience NVIDIA NIM on build.nvidia.com

Accelerated by NVIDIA DGX Cloud, build.nvidia.com is a model catalog with NIM microservices available through APIs and provides an environment for developers to try new software and models.