By Vinay Kumar, DigitalOcean Chief Product & Technology Officer
A seamless path for builders: Start building on build.nvidia.com, Deploy to DigitalOcean
The landscape of artificial intelligence has shifted from static models to dynamic, long-running agents. At DigitalOcean, our mission is to provide developers with a purpose-built, agentic, inference cloud for running AI in production—without the operational overhead or complex cost structures of traditional infrastructure.
Today, at NVIDIA GTC 2026, we are excited to announce a massive expansion of our inference capabilities in partnership with NVIDIA. We are moving beyond basic infrastructure; we are building an AI Factory designed specifically to support AI builders and power the next generation of autonomous agents.
DigitalOcean is rapidly becoming the dominant player and preferred deployment destination for agentic workflows. When the open-source agent OpenClaw (formerly Clawdbot) went viral, we recognized the market’s need for frictionless deployment.
In under 36 hours, we shipped a production-ready 1-Click Droplet to our Marketplace.
The results demonstrate our reach: OpenClaw has driven 43,000+ total deployments on DigitalOcean with over 11,000 active OpenClaw deployments in production today.
Builders aren’t just deploying models; they are utilizing our ecosystem, expanding into adjacent services like Backups, Snapshots, and Gradient AITM Serverless Inference to support their agentic workloads.
DigitalOcean and NVIDIA are also working together on NVIDIA NemoClaw, an open source stack that simplifies running OpenClaw always-on assistants, more safely, with a single command. The NVIDIA OpenShell runtime offers a secure environment to run autonomous agents and open source models, then deploy seamlessly to DigitalOcean.
Why is DigitalOcean uniquely positioned to win in this new marketplace? We are investing deeply to integrate traditional cloud primitives with our state-of-the-art Inference Hub and AI services with an unrivaled set of capabilities under one roof to support inference.
Traditional hyperscalers offer fragmented, overly complex environments. Our AI Factory removes that friction, giving builders a direct, simplified path to scale. Alongside cutting-edge GPUs, you get seamless access to the full suite of DigitalOcean products required to run a production business:
Managed Kubernetes (DOKS) for orchestration
VPC and Cloud Firewalls for uncompromising security
Load Balancers to effortlessly scale traffic
High performance NAS storage for faster inference and finetuning
Managed Databases like OpenSearch for handling massive vector datasets crucial for agent memory and retrieval
To power this AI Factory, we are launching Richmond, our newest data center engineered exclusively for AI.
Richmond features state-of-the-art NVIDIA HGXTM B300 systems.
It has a 400 Gbps non-blocking RDMA fabric to ensure the high-throughput data movement required by complex agentic workflows and massive-scale model deployments.
Surrounded by core cloud services including compute, network, storage, and Inference hub.
Through our deep ecosystem partnership with NVIDIA, we are integrating the world’s most advanced open models directly into the NVIDIA-native stack, deployed on DigitalOcean’s high-performance infrastructure.
Serverless Endpoints: Users can now consume DigitalOcean’s serverless endpoints directly on build.nvidia.com for elite open-weight models like GLM-5, Kimi-K2.5, and Minimax-M2.5.
NVIDIA Blackwell Performance: Running these state-of-the-art models on DigitalOcean’s Serverless Inference powered by NVIDIA Blackwell GPUs provided up to ~4X better performance compared to the NVIDIA HGX H200 (based on specific internal testing conditions).
NVIDIA Dynamo 1.0: We have adopted NVIDIA Dynamo 1.0 for scaling production inference. Customers running Dynamo with DigitalOcean Kubernetes are seeing massive performance gains and up to 3x lower inference cost on Hopper GPUs.
Through our integration with build.nvidia.com, AI and agent builders can now prototype using NVIDIA’s API catalog and high-performance models, before seamlessly deploying to DigitalOcean GPU Droplets. This unified workflow eliminates infrastructure friction, allowing builders to move from testing to a production-ready environment without rebuilding their stack.
We invite you to explore these capabilities today and begin shipping the next generation of AI applications.
We are also proud to announce that NVIDIA’s Nemotron 3 Nano model is now available on DigitalOcean.
This model brings elite efficiency to deep reasoning workflows.
Utilizing a sparse Mixture-of-Experts (MoE) architecture, it activates only 3.2B parameters per token.
This efficiency enables a massive 1-million-token context window, unlocking advanced agentic reasoning for complex, long-horizon tasks.
Join us at DigitalOcean Deploy in San Francisco to hear from innovators tackling the challenges of real-world inference, building with AI agents, and defining the future of the Agentic Era.


