On Monday, Jensen Huang walked onto the stage at the SAP Center in San Jose, home of the Sharks, and spent nearly three hours making a case that most enterprise leaders are still not ready to hear: the data center as you know it is already obsolete.

The replacement is what NVIDIA calls the "AI Factory." Not a place to store files or run applications, but infrastructure purpose-built to manufacture intelligence at industrial scale. And the numbers backing that claim are no longer theoretical. Huang raised NVIDIA's cumulative demand forecast for its Blackwell and Vera Rubin chip platforms to $1 trillion through 2027, doubled from the $500 billion figure he cited just months ago.

I was watching closely for what enterprise leaders should actually pay attention to beneath the spectacle (the keynote ended, naturally, with a Disney Olaf robot waddling on stage and singing robots performing a campfire song). Here are the five announcements that matter.

1. Tokens Are the New Commodity

The keynote opened with a video framing the token as the fundamental unit of modern AI. If that framing sounds abstract, think of it this way: every time an AI model generates a response, writes code, or analyzes a document, it produces tokens. Tokens are small units of text, code, or data. They are the output of the AI Factory in the same way that electricity is the output of a power plant.

This is not just a metaphor NVIDIA is playing with. It is the economic model they are building their entire hardware roadmap around.

The new Vera Rubin platform, shipping in the second half of 2026, produces tokens at five times the speed and one-tenth the cost of the current Blackwell Ultra generation. Fifty times more tokens per watt. These are not incremental improvements. They represent a step change in the economics of running AI at enterprise scale.

For business leaders, the metric to internalise here is token throughput per watt. It will increasingly determine the cost structure of every AI-powered product, service, and internal workflow your organisation runs. Just as cloud computing was once measured in storage and compute hours, AI infrastructure will be measured in how efficiently it converts energy into intelligence.

2. Your Structured Data Just Became Your Most Strategic AI Asset

This was the part of the keynote that will get the least attention on social media but may matter most to enterprise data teams.

Huang devoted significant time to what he called the "ground truth" of enterprise computing: structured data. The spreadsheets, SQL databases, data warehouses, and analytics platforms your business already operates on. His argument was pointed: generative AI can only be as trustworthy as the data it reasons over, and structured data is the foundation that makes AI outputs reliable.

NVIDIA announced two GPU-accelerated software libraries to make processing that data dramatically faster. cuDF handles structured data, accelerating engines like Apache Spark, Presto, DuckDB, and Polars. cuVS handles unstructured data, accelerating vector search for retrieval-augmented generation (RAG) and semantic search.

Diagram showing NVIDIA cuDF and cuVS accelerating enterprise data processing across structured and unstructured data platforms
NVIDIA's cuDF and cuVS libraries sit between the GPU layer and the data platforms enterprises already use. The play is not to replace your stack. It is to accelerate it.

The real-world results were hard to ignore. Snap, which serves nearly a billion active users, cut its daily data processing costs by 76% using cuDF on Google Cloud. The company now analyses 10 petabytes of data within a three-hour window. Nestle, working with IBM watsonx.data and NVIDIA GPUs, ran supply chain analytics five times faster at 83% lower cost.

The practical takeaway: if your organisation has been treating its data warehouses as legacy infrastructure, separate from your AI strategy, GTC 2026 is the signal to rethink that. The databases you already have are about to become the most important input to your AI systems.

3. AI Agents Are Going Enterprise-Ready

This was the biggest software story of the keynote.

If you have not been tracking the rise of OpenClaw, here is the short version: it is an open-source framework for building AI "agents" that do not just answer questions but take autonomous action. They can schedule tasks, write and execute code, manage files, spawn sub-agents, and work around the clock without supervision. It surpassed 100,000 GitHub stars in its first week and became, by some measures, the fastest-growing open source project in history.

Huang compared OpenClaw to what Windows did for personal computers: a standard operating environment that lets software (in this case, AI agents) run reliably. He stated flatly that every company in the world needs an OpenClaw strategy.

Jensen Huang presenting OpenClaw and NemoClaw at GTC 2026, with a large screen showing the agentic AI architecture
Huang called OpenClaw "the operating system for personal AI." NemoClaw is the enterprise security layer that makes deploying these agents in a corporate environment actually viable.

The problem, of course, is that autonomous agents with access to your files, your code, and your internal tools create enormous security and governance risks. That is where NVIDIA's contribution comes in.

NemoClaw is NVIDIA's enterprise security and privacy layer built on the OpenShell runtime. It wraps around OpenClaw agents and enforces policy-based guardrails over network access, file systems, and inference calls. Every request an agent makes is governed by declarative policy. Think of it as the difference between giving a brilliant contractor unrestricted access to your entire company network, versus giving them a controlled workspace with clear rules about what they can touch and what stays off limits.

NemoClaw can run on dedicated hardware (NVIDIA DGX Spark desktops, DGX Station, or even a GeForce RTX laptop), keeping sensitive agent operations local and off the cloud entirely. For enterprises that have been watching the agentic AI space with interest but holding back on deployment due to security concerns, this is the missing piece.

4. The Robotaxi Tipping Point

Huang has talked about autonomous vehicles at GTC for years. This year felt different. He declared that "the ChatGPT moment of self-driving cars has arrived," and the partnership announcements backed that up.

Four major automakers, BYD, Geely, Isuzu, and Nissan, are now building Level 4 autonomous vehicles (fully self-driving within defined operational areas) on NVIDIA's DRIVE Hyperion platform. Hyundai and Kia expanded their existing NVIDIA partnership to cover everything from L2+ driver assistance to L4 robotaxis through their Motional joint venture.

The Uber announcement was the headline grabber: NVIDIA-powered robotaxis launching in Los Angeles and San Francisco in early 2027, scaling to 28 cities across four continents by 2028. And Uber is not alone. Bolt, Grab, Lyft, and TIER IV are also building their autonomous fleets on DRIVE Hyperion.

NVIDIA DRIVE Hyperion platform architecture for Level 4 autonomous vehicles, showing the production-ready compute and sensor stack
NVIDIA DRIVE Hyperion provides a production-ready reference architecture for autonomous vehicles. Seven major automakers and multiple ride-hailing platforms are now building on it.

What makes this significant is not just the number of partners. It is the emerging network effect. More vehicles on the platform generate more driving data, which improves the models, which attracts more partners, which generates more data. If DRIVE Hyperion becomes the default standard for autonomous vehicles, in the same way CUDA became the default for GPU computing, NVIDIA will have locked in a position in physical AI that mirrors its dominance in the data centre.

For enterprise leaders outside the automotive industry, pay attention to the pattern, not just the product. Physical AI (robots, autonomous machines, embodied agents) is following the same adoption curve that digital AI followed three years ago. The question is not whether it arrives in your industry. It is when.

5. The Hardware Roadmap Does Not Plateau

The final takeaway is about trajectory. Vera Rubin is not the destination. It is the next step.

NVIDIA's computing roadmap slide showing Vera Rubin, Vera Rubin Ultra, and the Feynman architecture timeline through 2028
From Vera Rubin in 2026 to Feynman in 2028, NVIDIA is locking in an annual cadence of generational leaps. For enterprise budget planners, the capability curve only accelerates.

NVIDIA also revealed the Groq 3 LPU, the first chip from its $20 billion acquisition of Groq. This is a specialised processor designed to sit alongside GPUs and dramatically accelerate AI inference, making responses faster and cheaper. It ships in Q3 2026.

Beyond that, Huang previewed the Feynman architecture, targeted for 2028. Feynman includes a new CPU (Rosa, named after Rosalind Franklin), a next-generation LPU (LP40), BlueField-5 networking, and both copper and co-packaged optical scale-up. The claimed performance increase over current systems: 14x.

And in a move that drew audible reactions from the crowd, Huang announced that NVIDIA is taking its computing platform to space, with future Vera Rubin-based systems designed for orbital data centres.

The message for enterprise architects and budget planners is straightforward: do not plan for a plateau. The cost of inference is dropping by an order of magnitude every generation. Workloads that do not justify the compute spend today will be viable within 18 months. Build your AI strategy around a continuously improving cost curve, not a static snapshot of what is possible right now.

What This Means for Enterprise Leaders

GTC 2026 was not a product launch event. It was a thesis about where enterprise computing is going and how fast it is getting there.

The five threads are connected. Tokens as the unit of economic value. Structured data as the foundation of trustworthy AI. Agentic systems that act autonomously within policy guardrails. Physical AI extending intelligence from the screen into the real world. And a hardware roadmap that compresses the cost of all of this, relentlessly, generation over generation.

If your organisation is still treating AI as a series of isolated experiments, a chatbot here, a copilot there, GTC 2026 is the signal that the window for incremental thinking is closing. The companies that will capture the most value from this wave are the ones building integrated AI infrastructure now: connecting their data foundations to agentic systems, planning for token economics, and budgeting for a capability curve that only accelerates.

The AI Factory is not coming. It is here. The question is whether your organisation is still storing data or ready to start manufacturing intelligence.